var/home/core/zuul-output/0000755000175000017500000000000015155222525014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015155231401015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000203763615155231320020264 0ustar corecore2ikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ22|VVi.߷;U/;Yw?.y7W޾n^8 C4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;F}Zl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5a|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJBR_v'5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3?lm$K/$s_. WM]̍"W%`lO2-"ew@E=! I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG pHKZ{{Qo}i¿Xc\]e1e,5`te.5Hhao<[50wMUF􀍠PV?Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-Pk.@%=X#|ۡb1lKcj$׋bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGL`'žeVVޖ~;BLv[n|viPjbMeO?!hEfޮ])4 ?KN1o<]0Bg9lldXuT ʑ!Iu2ʌnB5*<^I^~G;Ja߄bHȌsK+D"̽E/"Ic3Y:1o@0TmBLi0lhѦ* _9[3L`I,|J @xS}NEij]Qexx*lJF#+L@-ՑQz֬]")JC])"K{v@`<ۃ7|qk" L+Y*Ha)j~pu7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWo/ݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧ ۣ,j|r6OSU;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!AzzGݑT5sTkOt]﮶j=}Yf&OC"msht: rB<SYi&It1!ʐDN a$0Y&Hv]9Zzz+]}%b'$^LJ<\HGbIqܢcZW+{jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'7W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎd 5s gÁbNW\: "HK<bdYND d)VpA@A i"j<鮗;"(MC`ʶϊGi+39#gNZYE:Q۷9muB`9`LDhs4Ǩ9S`~DkM{|D<˙ik; HD;;3!4 2Y.$Dwiu(lO:k$]ԜYLUҞ6EmH:=yݫ_o{ʳwΛA ԀZk"f`.,ל{=wh,,|8ݞ)Yghi= ^ ޮhd+xӁvGT.+-k)j_J>ɸv'IJ-tH{ "KFnLRH+H6Er$igsOKz%ViX~.735QE~qpYʕ{@}ZL! Z0!A촏q)[nf &E1K3i+`JG P/ EG whrH#< 1=5F8) 򃟤,OdW>UKĉ~b2KQ]hk6՛tcYͼ#$eoNoԦ=!&d.09DHN.>۞AK|s:.HDŽ">#%zNEt_LGiq'N~|<>7_l>9Hhs%y{#iUI[Gz/x;OuKRv'm;/~i5KI`5-'YݦD-w!{ Y򼤙&m^YAKC˴vҢ]+Xn#`iDf_%7K{)sH+"B4H7-xqz>xӾvs;mL?WƟ+fx5SukQqxST Ζ2%?T7<{a{N8[lr`$pZdK=3jwlL Eڲ t|*Eq G.tݻgqW4ŠqBFf8 9Fk Afq#/wu5/>HB)iZBBy/@W,v&- 6CBjxk9ᤉ ,As~5YޜZ4ΓVYf'h5~' SяtpA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUax|2*{_[IYJXFfQ* 0kq}mcEmLCEg=2ȴc>˚m^?E*-8nhױ1<=x-x:iLz+nyȚ[ՅwI1d[X]L4oI,ca%:g!|vwaBtx% BF%i&GjC6."69gxl!3.l D[MTo&r4ʻ8k~|9y3ę/tц}S@d{*W/C .2CBr6vKHɫ&edlBhwŦL&ٹO)X`W;<8L@D+dˠU@s[hiҕ|ڏG/G`' m5p|:USZ7I7/7cs9~}hLHq_R iE^1x5[O%3=-$h`EbDC;.j0X1dR? ^}Ԯե4NI ܣR;mu~+^k9ξNDͳ0*Vr'tdQu!4YhdqT nXeb|Ivż7>! {&~ĊL:}1*8&6f5 %>~R݄}WgѨ@OĹCuWai4AY!XH p7騋bZ%d>, >Df~=)(')r#{.1qZ|ZǏ\tr>Ah}ʪjnk?p ^}8"OM%Eޑ 5@f,|Ά(*(XHsc`% s/A[R$קQM17h\ED#O@>olJ/ť_i/ݼGw eQJipFr{u_py/]C 2i_e}L~5&lҬt񗽐0/λ{L[H* JzeMlTr &|RqkKyt1?Y˳Z%N柺hfӫŶ Au8Y4>w֫/&˸~%pq;% ?K~,/0'NZ׽KV羙OSJ&j0G[ PCbʍN^XS&E9ZS't$=tn̮&nu [w}Ab<~W4*@`t;6q6^9.EPHŽ{pN>`cZV yB8)"ۊB&{u:%lXZlvwohbL_#ǂsr_d >04SRm+0^PTi-"] O('@BKD6 {NmʐzRj.aQcb^CZ-uvpr CѐٱlGNzIveca=%1Qi F>wTLHUGӃ\sA֎Xpljlv ^tSȻ \cPwίwX"{>9V0ټ_`#U8VdTtD_GU9V ұ{q:ObUi7s )B ۊZlzIA4S#x,T3ѱ ԶJ=rs>Nb: Q6ˌ߉J%.Dl2ȱ%ܱ&6XƟ6qg(USok+Po$lwvmi8W_VT18V =| ub6QWCnY'"*aN08wuSEAVخ m3 o\` sHc# fqT .,ŀU|⦍߶/*~48âF,#[:y_YIpʼn)dk!J'Z5=r&; (y*b*O_ULT.ÔD[%s1,jЅ@k0Ցu֯dtKl$Y5O*GUڇvI`b0ο0~oI`b#FOf_$0!i rS/wvҍ%Eb/Ec|U9F-)L)ŘF`U:VK jeFrԋ7EDYpԽ.D\dNyj荊EEg]bÔF˩ք%EGƶ*NX)Hc(<|q@Oޯr^3>Uf1w;mCja:-1_k٘%VbZ˙#G6 `q+MPU~l!.?I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sRmA>d2UAkؖvlX܇Bz1U_#Xӫ+al H d\k/I,k,ρ|`zR/$@8VU^rcG"E7\qtS:ڝUyy >Vc11*?xYa8U`Jw/AcL~|;yj8TR#s"Q.ϊ/Yrx+u6*27fǪC%+A~*Zآ'ѭnۡ|< a1s\ T5҃FZh?EV"sd!@БU ^p%pO3|B5=2怕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7{!fRБBSۘ† Er/IGU}APQT]|XN X]FbKjKdO U6[3TTX)|*H'2U0:VunBl  `5/@ա06VNO8VGON@KgjyK?Wq1egI+ I.*F~L!Gf"LD&U 6tGd#fR*c ^tSLjnKS9 Ȼ \ >lr&}+̼d"I va,Jm_u)d靕َ| Vw85F3Liƙb<;dM-})C?Fw*IJ_3UG'+¨[9| >80\+ xJpΕ`p~mg˗%F Rg(6=/r+%a>w Ohght uЍaRs ^d6GXAf?V_mW puȇ S:tŴvŀU#-*mZ5k5r)_x*8ͼx@(k:_TX%[paRu~}#Ѥr %A%`;MxB[CzR怕#H% }8@*AM.SEhd,rKrʇ)br\+! s1CtӒNc_:F*`Nv;ogQFa2V%ZniE|nZ&-I,t*ώlo Lhnٓ'Xm R ˍ-~ά}hs\5TT%~am.>!LcoJrKmqvez܅E9t6FZXgsreHhlٷ+ [}r:̓?W~e6>0E8`Jq-(ed;W¨:Ä&]䒿e;0:|$Ȃ1L-%;Ƅ{dɱL;V[bp>!n&աIJX1$9;[?- й vRCxKVV+#lj@_RL;IQ8ŢΌXD@Z< (1ZRÜ:OUM/vư{'jYXE4S/8 7: `/ +G\ U>]B2/n2=8) B gJ3bcKo̹ʇ\B~Is 2sO/I!}xV&\b<9$4Nve^آ]$LGF@LjKٕyzH 31Հm-XıUXF|\A-2) ' RG6h?āUŔyj[j_ӂ~ яA弆^bDyzǖQ8`jXbsK?l58,?YP5䜭ve9YFznTEf3Ja\,@2,?WYؾNr<V` =V[oB5!Z\ļǪЎr8@*ucѡv\[|s L-+y{5K@dzp`r"mũɸHNd"yc Pu>x2;W`_VR<aӗ&D<=h-Rר|/r _ǖھcߖ]G@Ն;UQG1 '3Jە Q88ASUȿ!:WѥLf21;d9OU᧯MR3V:<}xXh//T+coY5Ȧ4/m0NE(G2[+G~H'5ipӘ͏O +Px SPp.,?Uvwmh~ %˗eg#*F-J)zslue۳CyyL"cUM|h Ne <xRcUedk;V3>/mP=HcUf.ZjE"(LM&;ͱ@ƪ.ߏo\]Ub|,Xҟ.q< Q2=fLj>zLBY(o~ اE|;Ns$9yhay;.);1z]Rvٮ,"&leM#p^_+c/=`4_WF̕1 +.zy,mH+b*#h 4s퀅Zs("_Ӛir4,q^{@V"//rAs5ٷ˶==>d0*0-_nCE1l h`>]t{I^wKxef,?05]68v?Gy*t{uൺd?2$7_~ϦSS7+ݼ2+ݹҽ+fOnp 쵣͓?: zԂR#A4a1`7*OtL:.s}~z'd$x/}d1޾1>tӵ]!ODCkT=?8-8Z FS$a /E&ώ*ѨL/aȾ׃'ESXjG̉̀3oPt0)7eMwP#48P@fAz]zqŴʃÁe퀡,㲆PkEwRe=c庎8̖ϙ[M, 7la[/݈{fgf~6Or_$i y {]W4|=(Q˯7vؚL@$3?ZYuSyi`Kq`=WVb--BZbD'fM+AyTj2 <E^|)I( `J2|5"&(E UNE$ >9A.c1r'PXA{}ea~L'RNu]T?nլu5kG Ϸ$U{#$I c13, RAZinH-l߄yUY2<'X_p:#N`)| b%/q/]uw~uVh`p@!8! b zb.gYsQoydxsb%o' %<<ylt-'ݕ<4&GSvpJ "~bA#z&0+WrƕvѩP tK;Lڭ@mÈ#/e]|CV&:\u٠;ix=ݰ2I'1Zd)m6N޽l|PK(e<חU$7nn%q@U|QH&Ϩ7W'jQ$1x OҸ]nQF qКB c5?<{S`;$F!R,O%v$ZX$¥#  P7oE HCuO.VׇBf ɳ+͇w>`CA BsYxDMSlB~)ͽ;?U^.dp}#eK=Je [ )obP7{s4(jR$DJe4%?Zf5y-H83b"K91BXY=;G'aX !u2$(Cևm#X5@ڇ7$ Ӌ!3‰:t5H X7{9:Ou)Ezbs?$01ģyŰ`eɭ_pR^Y95& s xܴDۏ;e'H{LϿ\vS!'Mx`4 g^Y f 6`"A!l  k6XEg$tХ>%/߿{!>*F$m]ͳ oD)/ )ib.dF͠M<\קRH}$VN῅ohٛLq袽rÈozЉWαOMa@i*%I~#C*I^F4M0nDK#)k'RߢW,6$L! 4FQRo^[0]h-Zѽ+b980wAY%[Ma=Ӂq-W vKm;OMZ?[9cݎ)?/9,(J3Y ]%~clW d*D#re3+lo͇p-}hATie= .q1QJs3,iWwe. 8$ -i.lc7x/Z>ms7!d`2,6dK65[%:~SV$Vl{'jIܒ̃ }(hH}հFt {GbMy kZXu̝0^'"N\] nJեȪ{睜r,# AџA0(ҁ65 &zbYp6޷I6ϳk {:G$®!D91[*X݈BCْWcW{M5Q=L) ]kQ)nd>I4['#ySk) bԡH +W.D\~7µ42`yvG,3L3hU8C˴8je\ȤbrR0װC #5S-%՗VUX>gIUl[F wVycT r]mZyp:8-Ƚ] ']S;]zO1ҨnPKe!&L"nn"*"خ{R 0]gf,nsm(x*~k`ٳw*JF09zoJ݁z@Ӟm_Q6Q<{ : D WyOh;mn( ͵&TMOM6Jj (ہ(a5,e"m~46ߑr[xGdAb\zxSSn]qEDPiwjXn.߿jm;C6` 5)ڄK55&! ؇0 SI<#YZS&D+fgX|,4,GEPhȔ 6X?4&+@ m!8P ao?K , 2˻OMSi/>[!LAmS7UfJkj8|j Ѷ|E'rXJSKOgI^]qr =gv`Yk2ڑ$b&ӟA)BQMג2J=F>S pGC<ASR^au'G݈VV~.gyPhKkkV aе-ϳi‹"2aC+c느N\b|O&[C&sv"KkxOP[f%9`\}8-…%ԫ$\P"ڰv΍ R";6*dţCZ%whHLE␵\`!&=֮鵿_z_v0ilת}.Fu)[1P[OIޑXSɹJER!7mU<~wDѕtkR"1]Uֶl< I*[꯳ﳮ˴TG6Ue;yS[ <-V rw 2r&IHܹ\1૝2*g|dqbBhGA"UƳG~׹C<P*Xi+0Oqc,O3DpSar.iG9p+sư0g]%yoҬNe\fWwY#RT.^z )cSWXTW־agad7ToT-v=܁XX[ψv{T[H7Wr ΢>]˻t"O7ʘa1tI ru*M"G uUlu f9AWg`p=Ϙŭh ,@:{Ҹ@jl9g67M.&Nl+~H]m ٶ}%Nkfg (lB'/UQH _s9q $I0PG|>a1(nC% 0[9fZ%1_%;Ԓ@f;5I1s λ3"bXwӲCH*@:[a{(8hws\~x?KX8aR'*0*v{ y`gh5={e6T:\_(qb3l( ͑BʸkTO/HG =:-DQ-rRiz )ȧتe4b+9(=:G_:ZKTrY(ށtfչwu1r0D8dHw<LPxG1L(:Nf͟uzI2D} ck^鸮T}G=~Nf6c1Tj\CjJ>?ZuSCWo/ԆO33՚wi|^C>'u?-( 3CGN3ڤqu} <6nM { ,(O,+]:g)(QB.|7W(Dee%ꇷ%b @]$A$VqVnл;J&`= BPdnz$j x'yjWNaqu6#>!(H ;QX&7;o/׿\ֵ,ߙQ$\E_g5Q(rS=%&\3Ž!Cx\!%zK5I 0>s;Tޜq%Pǻ2mZZQvfHb/{w?$R{gNYdHրPܺJfK;dR$*WW_D(GdQ?cTΔ.9}1In֝-O*%Kr8O8+z})= 8s[Hƃz%=ޏHygvPYxu*m:E2n%Q5NJ *^),tާ`VF/Rx~4P;ّ6:&Jw%m>>Hz=I !EyAؓ[LOK {J']om䮓`Hgr(DQO⳥^3]0M:^JF90R*5)/shv .NLr PB#.AbK&ϫ*+U̒jN_Z9%Хt%x*pǪ~j @*/;&rRr>iHIC+dǃE¸'ZPDHQj5ҲIE 0+1R/ؖɛoƪ힫{Exja{2KrB;C3\z4]^3x<\=GfoiS±<`y5G%iUe(+dVh-Gƶ) ^ BӢEDmdz':O.=OsCّQa9I4&g,VsF_@J^U8ފ֕&[,>F=%wIi^V /Gzx$)1^MQf35?&i2s\Y暖crRѹgu\Xe`L|=Wlniİ%c2Og ax*j5F\6(ڊ)L x/)Ӎ#ΛOoGQ gT8^q(~~uqR^~vϋ]_?O54 T. rxkZXYHY.B? 0dyJBK`$ˇK/lk<-`kVq6憕~ehVG 0%58K|3V ncyܪ+Vua;ew?px*hfm/bؒG,Q/ e넲-eG(ۑP m3lAeU밲n5 cvft:| bǑwoANz#ۑPo B:b B*v$TlAN#ߑP B:r B*w$TnAhNh#4ؑ` BuB- '4|ᎄ[mAh=v$4@.ulVz!iy=%5ݭ?6./0܂iy>.Dpa⪨ MC#\IzQp*֎ J2jT9Aee`=2B{Hobۿ  |:|PÁ}bW) ɜBfUQb2Έ` pX>|(h0ͨFW"dvce`h;N␽YU{X&w>mi_?~ TE; ZuAkQvk8-Ǩr\xYFY?ɭ|]t dٰ=9@!Zz$xxrõ QXHHoU_u[>DcRZ+;,W=] s`/Q7]a2Xf-m7:m~j:*4qhpwάT?׻ub<VHg S:e>ɫdH|z\Y>0+Opˬ/1WGeCa]k\e6 d OK EװaRWC<:~S%iH/bz㤘h:|a @ՏC=@j_W5 A;e£!|k6|z7$*Dj=inyR*-r)L-){xS ^0qc魆tofb6<*"-&lFr\4x0g*0;f׷׳Fp@|#b^hnpm挊"AdIc~`gbʔ%THaBmڂCN%d~qSt5ʥ}X({8Tz˪⠒\a>OCw7Dr=W~:xHѴ@^.۠.KOUa tLqugmPfd=mY 0(h̚HitH= /S' k|Xd-/<YaW VdCmd +,J_tUeJP%hYʊ5>nԺDV<-5 <6]1.8_,ɂ0H ],ѡJHWbsjrkb4~EW|7Rn*H;K 07lJN,`'$gpK}&ep}60R(~z_%ogv+f* QیƋA#N0> zH/yW9_ܧy}U>Im~_n!Xh"tZA?M:W77Z$gVAL@c RљkVnẬ0#50Աo\?AǭYiȺ`責Ҽ  !MG`<΂'P: +8v8zzT١2iaN0~Hg!lv8M{Eь1h*ƠNbr&kK8 ե) C5@.dYL/]վen1CSe1>oyp7?>#\6M!IiOzy]P 0 Ҙ\[9l1Fk>!vX!juV;I{)D7gM;tuSw P{@~rYg8X.Å{j0WӁ5,+Pƞ0]Y/vO,.\^n8@_d<`jKDf2cS#?R>>]ؿf'{Ϯ&a}2?'ip#xlynNwm #3yĪCYa $C•BE +H=Z6p"nM8,N$JWLJRZwvǨ86-s:(֞&QT AȬKe&!.=l&#Sj$7mO':ј:9M`@h3XzW{Yڕ!>&e9#f6?\;^[ wFk$]z3@ʾC©[CpTGjcA%NHc+aa& G^%?"yA ?d俧wem%I_Y1}14`0 *D*l5'xKQ`g\(Ȓ Ff'zz8|=Ϳf|xُ?~.c[L~NqG|o7]9i&Ϳz_oa< 7|܃rY?PĜilXcu^ VJ3ΫKܼoaZ@'L{W }fJ̼l=^K!8Pg5yv1sn,vѹj: Q(N4X2S#ö=v]ctP#ǧ"&%-A T]{[/OSTщyyg#BϷ jf>yE F TN|9I8bwMvcmmnXqх*՝ M\T8!b0y#AK2s4)FXm҂Ai05W y1v) |ʋ.˱Ŋ\4XV5!Xf>q #ϰ[QA=+$:0"&!E`tF1Ag1l@1l7(`8@hWxu*T8(0)Ԙi'" ?8Wh1H*Z 2FڛaHtTy68UVկ Z(E@ Ge-5Y킟{J]) ht\xAÏ"`Y'$ّ Z%>,^ M KMfY`Lƾ?cީ.+8N( %ֶ@FTX gMKj֢fBGàhB0K/BaKtqxkllZeh$,>LSSļZ!^47w/|*85d*(A l7O<>fJހp{]abXů$]I/5vQSa`br;k}OE78ֽ6z@ &fĺ%ޭvVۤfmj?Z(ȏ> ʌM*&j r FiF j c&Fhs=llK8>p Bx:%t oY}Ϊ b,>+E0%rt $YZI31R &8q]r.x1p K7XN݇Oc Q,LL]`F5lfit_>P$+9v6嶭\կ`f#@|f$^I~ZmIGMlpd3SokU{vsI/k"EvȮU`!:W0fP)p"GJ)"iUWqtnpF ه&1nzee*Qa$AbLd3 Ee7ʙHrGf5|xJmQ&OQw+jM1rt[i M(EYK ?Xz0,~zdƎc M(1k>:bp9 E҅}q{遣e?HjaX=wQWehE5+D6U()R|;{3 ٘.<vn ~.V6 xnD1hSxFTr!Tn̹S6<<.˶Dw nKYF)=e~Clw7*mXI_9։.88j,Ƽ Jmj52!cI"/v_ouܚm3r+ ߻kpU(oupV=b"SK 9&K<I8~V/w]pcd)Q4A$*B+Y# !Lܻ@쬪GI%m~0AQεht>Q!1(H CĻq=6+$U{+ "Eb66ءƱߓnoXeJsEt>mżR *m7w+ *:a{}` -άŚ 8}U1cMtqYru|FB\F"ұ/w=p*H>H/pLw.86)ArtB[' Oɘ<\S$}?8:JI(jQ@-B_O0E+؟)ԬSL 86z<57nXɲcGcFJ3 BRq9b8Z N ÷NΓvY O[)q{8zq0kTcJ_[P"En)")e+y7۶ b0ya 6%V#r]cdMP$]܄ٿtIa)Xxtmڮ@EZ+11mZvt gpPR(5gyMObʷ0/2ˊҺ(zv{!v}kML^)eG=8D-* pMKBsʫG_#EY>u9Z΋o:ZA( Z͊DIasG܋mY6vvpεZߧD<x2g݊i3*Q;QD B$~׬k Q;vjlETOœ5:.'T 'yTגR*m2Wyuv΋wr򡽻>5+R1J=/=4sĵdۓ>f [nK삣,FׂO"X xB`cKvH}_R?{&FF Х4и**o ~ܣEHLv[Wp]"7! ,{nU} B7dsm;i$oX.e=>SHB[mx=y}Nz(| Hy*eڗC.p 5d *) |/.I?8o<UJ jɵ!ɂfT KJ|~^vQYMPz opsQvn>4bhݻ;wLGZ.&hu5swϤr&N=ֶ`{fTm({^wjhlry+3 ΁]{vvg$\vI˔@*DppO`va7aK嬚mO͢N۾t/wte7\7x4`zj*1 \ZL9Q}"xp +xrO))4ki-EM(} y-{-~ p,];x\/OnZ 6/VDsg8xjh/h) #ɤ HJuYuNy oO޹Gf5r (c.ͱڄi쏒La0 Oӈ4C gU@]%}IVʯ6]p;%k#P~|PF1q3:N`)'!vOituo٭Qqw5D qrX1cVx݃U(i=^.[\p5뻽Y, [plkVb2f}ǚ4C>NmFw[ʓ׀ՖVb_X/ACׄ-A_/W!&X+Ca 7#Cx?B qʓ.>q>,3eh<V X1 %ag=w9뽇Wnj>LB7RlEߘ侙swŻ&JVdo'.Bc4Oٻ`8Tތ;.|EhԬ.~7ƙgB3?FBy{g~h G L'|?9xh޻OE3uD#Dy u/Շ:mRnnfJMU6@K֍ނ0?nPۥ%Vu6\<6TxٟEX޿c0{ )?^{z|3N'?ɝ\鏯N|z?aOI¤ƁFM.T(7\ OPG'OjP$ 38?>ˇ7i=޻oG''Q,<KAŦ<$qU'|Rhc9~m?g '?A&I2 UcE E1 Rz2D,ysk be~\8%o|W^}g3w ^ќ*z:0T? 10vGiRpJ03*W>Du*?7 I U@ReQh䪒qh'yY!C>)OLǨ> 1t洶J>fV>\'Wfm֤^C@{gƦD,&Nf3bӿn44NO,#~u9W"w["WFW3I+u9y[3FaߦaFf~t/Җ͙3MaNM9 [qZ62--S/3+g|b֌z'L~ғN2F'k2"[8?r.KKIo')r^{3') Lӳ%gV'^|&+Z9c5EY4,o{,Dߖ2g`-@TbffE^IT 3 in1s`  &=Nxˌ)ys$O&OT2) w%]pZ1Kll=]@]b)`RQw* BtJ^v_:g*|PoKܠ%@@Le|d:@b-Ek);Zr-ovh{Ǭ &G F+~9QSYw^N_ 7YK' ' g~\J;,pBʁ$n7՞?ዹ6Ng,}2CɪمK;G,VnDP˩N"؟}7 ]ݽv@T]↏x^OIޝu=⢤ 0K5"De dG ಸGHytUᱼ44աl3{Y3L(̓RT[) NJ wX𴡦Z*VVZ V{{2ʏCk-½Pnkh]m&.o?,%RaeWQE;VjrT+s|FuYQ㜳9]'ǃ*~K8bwH)/),>Z*y=!ʪl=C#1X#%,Uu*^.5QGglgE8z2Ƴ"[1ml]cN(4g٢d lO@.ܗ^ꩧH=4JJvaZgfvu-u:* VKj)3%x:x0x ˷>;s3XQ;KBg^:x$ ۹8UT# pygPc:Kyݭ?EII ekǣ~fG$7P?Fa &0PFRlP<8G,Z'JSWr 璄; ?uZ^lXlZQqGlǙ!p3Cm5.,o2k\jZ;]/z1ZxR݋`F -pdrX&|JkyJtʱEVX{a|zm/nZQ#/nZӽ+"p^l(o?ʌ=[JF1IJѠ 1@ de s,2)񦊵=iE_DhNo>N~:oM}7s̚ 7#@ 5o}~Y \߅Pr9nenst5< |>yWqNp.5N d!n[iT~km|}Ig[om}=M6>}uL.a6a*5"XPTJsXjhf^b<Wd0K:Fq7? ?UUdO5aozQ'Lۇr sniM/]y:̫*MM&Ֆ02誱:6]4p)  HB_a{#[o~eUuj!%LH=3p 匾mjcJtj?(ڙKǡՆb Fo:1oB۶ٻk޽(Oc1z:UHA >6 yYE@Qdkg٢_Hߗ/ܧfK*B6L|zw~:?r8sQ5) ) VZ gRMP/Oo"M&yV 6C)‡ֶXPvpX;ndD^ᲉD2[9#7OSB̦XReE~i(q:Ub',bAطLRav_wu>sلG%Ab>*Q);|B3f@C~tSaL"' QS) 9AڀkC6T"{G7 A<թvA )&Xd`To9!y%VQ'1f1%r%qbD™RPUH4MjMD\1/vsIhx6 01[>IS7(&*;c E67&ħ jAf@%Ը&I_wG=e ;j_6鄚U 4W[a:t{jK3;٭q!F^ "&MJA3Xal1$ d\Z(`FaWCut7ʺ}2гu.Eȥ9R%¦<I5XкH3o>?z`@bg:D<-PG0U;w%m ykl l6 9c=ZRr5_DvGsI ["ïC+϶\`(RY #N0׹@sr* riH1Lv5v LE2c@뼜,%ٜ^FB,-"R6+ ɥCR2"z5n;uFo[@x~ϼqّaS]5%!Mf]ޭ*@}d \}\6rttmj*}T.xe2#k2vՐY,lha$DAhpc?%L6ao1MVoQ CU 2Fs' y=#V'53TgZhe H`3h B.Y9jT$ {ا((+-Dtk5iLUZ p],n% `uAj4H]AJeZXIIZ̤Uĵkھd` FjGC=s{ϯ`3)dکԉ2Ŕ€5b\eES~JK ̥sTr6M $S^SCJ*?IVZXK_`R #T$UaERiRamdj- tW * QTv:*%*ia0"͞7):~ !WG wh]_Ig9E4@NH0mNmeT,TijKHAfBKp$\a:&ML!Om>_vL bLȡ]*gW4k8˂q >OXHZnPjqC|%Rq)ֱ3ipM{LjNKH6Ii\AV4Rh}`y؁1ݖѪMNZ"@1kUц6Z nkH 1RQnjM.O)H6$E@锍CF4DE9g]yF|f݆0y+`J*M~ D[Z%% 6v qSDҚ5CLȟC__ezzfamEy$L锌޾b&ۖm0nv^hҵIhҮ5BFn#D ZWlvP ~2 H0Tw`b4"rL/Jɕ+*|RF7 MI&4/Y:hRo>}E-|0//';\P4WhjK_as4^oa t9ˋӓq}r7+`c:HގX_j)O?jvwϵs:%V/u_SyeZ"֤3/Un";vެ(^_TEqs @|A ^X&&'ުu{RfX5?!+q= 7j,^A (s9 J#HA, o{IB0c@m8dE݄ ,? QQ0Z =mB7t Z-ZoF,a`#4L[aw3hKQ`LG9LGlcr9wh}{X#6O>rՑE}es˫%om1T&)O檵(GWC k-%o;ۯǤ2qd턒H΋>Ձ `BwCE2yqZ^ o |3?=y]wnWh˕+g})vG?}~9 UƇM|KoP+GBc ] G WQFu{O%bsӚ3 5)ҫY$B~ڰ`X7t`9gj~'_[>oLŇ+u 0ކ 6#-cv XGOUpa{`zfnWSs'Mp QRyy2!NFST;ӓ/=OgrwtLZ5Bc0r/"B~fyfcw[O F ѭݞ'z Op}cE,ALNi6]bJO/e8G9~;?s2Wɭ7eZ@$yB{qPI/w>?9g寿Mu9:Hb? b-O?/E5ޛ0w@N>_S1w{? wwD9>\wW4=36 uA~=unÀ$,=ݫH~A(Az~q~svq>Cg u^)R#^tpzq5Epr:k p-_Wjَ'6.38}e Rբ k7h'm5LuĻy =Ea4=sz:&n%/tH7ⵚ\>ޯ{mr|.q2t_V7H 2qҙlqDs|Ijxq\};OO.OsD3'M&w$| xy|G;*0QeiW8?R ;,mDj]rs L0*~5"v5 \nV1äh!Ú:(54*B&T{O};6z40~Xy~y^\^Z>~*vPf__`@||N 4ih9W":i]`j4X`Ծ5y}Msyxmh/hB/r}ж=?K3hssG;W2, ʢX)._z~^ +\DYilƔFe۝S*=CEPH L!u'$G$•q[N\`*.kxEPfVVR+]Ӵ3i;1S ̞'ÏhWOhDhFG?ގs}9iLZxyyZ)_Gq3}@Wn2_O+M{oSs9MؼJrƂ(Ar棂ruPIh:ԩ048p)Ө}{Im:ԡ22 Jhz^piNQ3 ~@6~MMo3yDnQ7h3 چɮuӚj/i~Nn2haYfPF]S7s7ƄKpw#ژ$Wװ_O[(kHI Lߦt{rh+@kY99QsM.˵\+(jr]8mk;N1oǭݧN Qܹa ?D ?6wET~?d ?d͏.#6~MCIvSR+0R4 DbjST&*vbFZ: .ɉ vfvOe0ʧ~*C%D>T+7QahJq&n. h5]ޘ@EvL><ٿ퍇yR @  4\Q9mt\Z%ˬdUI9p]E~Na8EɞfdmDY:#C\?~q]x>J߇|$5'8= ZcJdV1u}Ia?,~RgWWx =Gz-㲑^jB 5}=@ (J25rsYS@Ƴ8<9kBj u]=7ȤYSݩtΞ' DێqC8E;tʞaіF,;~ӂ~(DXժû'k?#ԎU1 zMJ ܁>}!61p6hԑ[VN+5귝QedXNg潯6ZV}HnY)+jUEvp1ו!͢[OM ?C.7F? )q F 3',s{m<> ?bDM1fT9$ctGQ>mb<&4d̥qI| i.VӌmBWۍbګiyq5͘YGgٝ{?jLn7hvT7횧ϓhZM8߷`XeZԭ?C4֞!Z9xROiO1#&$iM.5=QN4^+w^H(Š&!ˏQWΕgvivIbv Jh~}撦j깽:Me#gL{Lh1M?iFcN蘦$Gn˵D[/⾸0~ k%a}.go^r^^$Vr:.֌qgݘYӖ3 <i1$Vj%A{J\ٛtseo=deI5me/BшWk L,h8ci3 G7\8naѢ0hPb3ݜI({F<#kĭ3@XbO;rz`\*$&CLZ?|·j>)w Ur7Fpl6n)'b2ۨid% =Nd#5WH]G S裻LG>kpE1Z :jA(;XRZȭ2'pq,&3]Ҵ<܏@lT^_ܼe'Sԕץ6ӦP ?5j|-F|R6)^^isAݭ.@=~ީ{q#ULA{n:m1z_ki X^SV5_w_vun\YUO߽tėD{7{; ǡN|_^s7Jh`T,ڹ02\SaJ/rʡ֕R7*9v4r;kRb*wA=Qa#PbI :,)K_wlulX֬lPǨ 3L6>q!hR!Iϛ~] (R 9,cr$1L 7tX+ڞ c:3] -UO |9C:\꧑V(eeOӞ;FWнO=|QuO :ɹ1eWkN'h ]3"8`!0\{-$C;%g~ce F SG",߮8s91m|ctC9{GԵ ?=V/k_/xk{OϲP(\ckWc> e(kz~(#ƫr(۲=U6aX2)7$A#X&4,q* FfI4fb@r]r]"Ibs=2p028V1btBd4 m\/N㐑ـ4W&W=LoT}%gV~_?jXy=tQT7wmhj $r hL024j}qj`Tfr'TN)C=@՜?T9\х;fgS *#TgS^͘kw0N8TȹTPDͲ\e/?$({*ʨdͲld/SfcqF0(Ͳd/#C/Hgx$eٛe^@IbP%vPdٛe9^Kzi0c|`jͲ76,@ifRDeo*{kt*fUY=\+,{H"q} exHÕi^Ow=eCG#)ӼrJJƖJfSg%o(Fl{/Hu>u.4AJH􁔉JSC5=[הcYzQ1*YrYMozF)іzP*8ae՛U٩^h='bYf{й~"?|Pz=O TQSN)_j|HF7_ l7>bIW@/^~8$|Syxuq-pyBK)6EU*-|ֻ.?ޝm\G݋?~CԞ_lz?/7+o}J>^|ꢸ׻ G uin - .pko6X$E6yx}Y| b|o>~ )қ>R TQ%9Zp *_ջ)b|C*(8bGqTHBT.m%Xb`*B'j,ZlJd\Q!I5Ζ@ On** 8hŢA0B*6f)ZypC;I6.?7*eQ]|/7ӨZ-t|&2%9mԖj]jT남6=p ,XS5I!\־h HmMa/Za(% +0 %j|V;V|^6:K`׮8[@О^~"xVhZjli5γPRjAW$.uo'Rgf*\3uw" W_풴PE/ 8U3!U=֖Oa cnV 2ȅ/@a FCaPrػ7 ^4?ӽT˟Ol}Da/68+*&$#! /TI9DGl.wL|]AsFfoh7TU)aviFCC"0ĞGev;0H kE݅ ) Kxl`c7i֡M)mޤ9;Şŗ"T+@ߊT*6X*f⾂჊Q%`q)D1ld6DAC22d0L!  j! Ǩ)ĠG2 ĭzy'UNؑ\4sr\InFeגFfmQ== #_Dlia#r%;T#|ULSje . 8\I'dd>\"Ki]d\2A%؏;4%HH02.42g]؋_} ͔0o+{*3#3{,V^:JsSTZ/MɆD,s~J_+6LMQ4b|]4T"&xN.h.AF%ܠV*[f],\۪ɛY0,zU{|& 8S@=$#F2m]/'M/`9Ma#59m ~Mx^5zk>@%jY ̆4u9-f{Z i+䴘j;A] ht#O`#a. PuGSO}J%hArJjl0 h;a*sA =}`&.Tiqѐ1́?/|Hx.G^d_HU^ڏ #t÷ۻᨇPs]~`bDI$3hd6$5;t#WE:tSTec(17 F72曂YA*7/O 88`Ѩ.U@p ŏoy?]e5]Λe,K0&uW8dBF w wjLU߮J+lc`9  9u:N q[VQVQ"<".@ :uO̦m`lai4Fb9kPzVJ8:ṁ9g%!_x%MS`/^_$ĴKp 'FfCl9rХ{j29kϕEJf] EpS&BºKnn uL\7.ƃg8Xc; ,JCj|*nS`EtE!/ $i!Kp{Hs$2ld6D tN_#F[`LGmPKXW"Me)# Wn pX+a^J.fړEsx"Πru(AM M)sj}JjL)``)aNR&%+\`nZf|8-tak9ee|@. .J2*|c3>1"4BR$,8#lHC[8)V֠ Gbl&!Jlmk*: ̇ )-vZ̹{Vc4&'@XRRFِ1&@2KIcOs!UyÂl \X1hz1J+4ld6ZܵNβ|(KֿbEx(Ƀr5.6b$0#-,Ƥ|"olAK&7O $ XdF񰮞FeYClhi(X+MhcSCD!FC!.wآCJUVD:dҡtBp:B@b62A.5ꨡPd2yY:P;A?C%)2ld6¨rm:x'm"`1rW1 _ k`BE43ld>T6S U=RR@ˆ)O8jTDS"|PLDQlb_D1:Qh䟠 *)D1llDF$׏?قߋ\7y[> uRŏ/7_|TmI+p?6r$;3`X%Qi_ve'~EJGGmn=bGXBpUPzTT7AxlU REHKŢ"Zu Za,D8Õgw!_7\le!˚r 2ΆBS)zYL2vYY-Q\Axv)Z~0wt~@!1Rn;ż;22>gC'X$ŰG!R PwRk@!574z2h44>]ZIƂ e5~QK"׆uQK`h)؎Ռ 6r>N`'$4ם8y`ɯ iA) ͅ+kbܦDYj:yl7n1$fb# I*)6hq>"1G(y TuQ!XH0Xl T8 BƑ5D\#[(b#p0(Gvmm0uA#ӳ9sGN%):G6hR*Za(ͅH/HdʨP6%…]bw)vwN:*  9SI/ׅjnp›..TJoy5ą 9]8ͳ$ EGw`)1AQAX FH] :S$|S.@`J |NBP!@w庴QZmKq])D9ݚ?3̄.PT^KcK:^̪7(7TJFؾŊɄHΈ&JiJrRV*!K+:Bk „Jc!g+p/W l&{1yLmdEIC !/w&֝C?shFTlK܂:!\84E@.ءʒwp߁XnT ΃@* JJN wJ0xWP[T]SBHxrsS|̠Ri7~\0L)pEŖ25׊xrs!gL0}N^T5c0Zb W⿇RlJ%4v=>ۂղmG`hD,HN1K[.[K&a]oľN'u{saN[;X`DX3!EܙQsN =_ޮXi8f"dv1DhPa</b\-Cff&狸0s'i̧?km~q*G h eMv\b GE)s<:پρʞXjf!0OVE2Rѹ{\vX/V20>(JbDaJc)iÂtE%ӛ.vt]ܛ. <9BɅEP'cϰheG+D+xkY=} {A+O'K<;w;w4dǔPH4iRk6uH;aED[9wrxH*F6pY1T_PPCC@Hq8fۀB0vHi&˙6Z.pvw&o8,5ex": ܁z'`t@bc(H|Bqr U]ppƚt{_ e~Q.؂AyK9gV@&vlK;^K:Pˮ,0R={(؎^]0,Z=i9e 2JL!ځ+HQ*+"xVunJXbL.V(!g r.YHa,hJ(>C[N&KdrHRIj #5:aFyfR$A N2ōAs2I1$=!#RNF1=Er0u6KҠr`(9 {[z(K9h$h3Ċ`xMDCz0ayFN< $F'uu)b |%) $XlcKI>#C1[#H5nYbW X;yKTWL/k yFڭvki ,P!NsGX؎v)F,tB՜LJbv0킑QF(K0p*V( F>"3A-E(*eE/x󨳀k'l5轮>_=9YSޕ/M Ǡ׏f2dEW9~xŞ+7՝3XDზ]m{f^Tz][pO :}8۬ X>Ɉ;pD0mzD;>rJ\Y">JFZ ]$p  ;vl?\c=lQ-''[ .'[aWl{Xzn'rOCXPp)Trhu9ϨԊH,':n[s( &Kuw8x\3inF9 UOPy?i.]C;mg|הl}҃>Ҥ~ZsΎqo2l<4^ov,<a7R"vҮ(kڟL( 0[$UQ<Ϲq$U60O1J41 ~?VEGLo'§ҋ >50e5nkDbR ^[ZZl}*r/n:3 }Ue'#k%KajשG6xل-z]-z^>𽋽 ʧۨi?LFE 9魝bJw ;5HwQJKegTh Kxo So% ]+x.F:r}`h&gLOET=qq+ G<I|v.WSq^KIRْZ>MSc><*fQaxwm-^~bYط>$<׸_MMq&< z5b̓KM'|?g Vpw.ƯAF@S>q#pIӜ>uALjLzUpudyFЭQQWUʟB Bw­bnEc~qBNJ.S`4l:O rڅX)&y.s3XɆ&֦nfS3.ܖr1G4]l~$yIijɧqT΁6|='f<Q).> !_STBL0G %ǎj%FyvthIQ-PTR-/Jpoy5gCW;s^e՝~AiEC-FzCt\,R'ѫwʇL[tk=};KA?[/"~2.EuK-̭ c3 :'(%q}}]m7#{ F)T4O9%-]SuyG|Cphì*͗/`u TnjFWRA)qJb\x[Τ_'_Vh*a%8reum{rÕF7)6h)~+ѫ.ߨ٫ly3,2 B/+fC-4^C7Ȫ9Ϟ1ш=?>H߁v0PJ-f$X$L,NPDR}nz?D"pK6-{ n ^VsՀL~^ˌ )B"̪< 7(цSIEKD±A\ b8ZqAj͵U`=1 RDbP!6)d6 &I2f0%ʄI;qv+{?TYƤQ49Xg.3- -!&G f)0PyPZFD1Jk`D!(͠h3f1jdZ6yb45'B&D<(-VQ)kkv>J-9i*dH܂Șin13\kI0- n2`KM"d3C3 3Р$ϳħ`)KU:Oc#Dhh 4#^.J*LASpa u:D!{V]!pT1iMoi*󷬂hSGQ%cNBk`O/Ss.7wJcWU˝=c4%JIIjncP9$uNF,QsZtjRhK {eߠ'16.%yI催!ڨjiH DH,!8dk GԨ*;uRl U} ` ɨ\S>Ec=uk,Zad1!KuXXteݠTGhOM6TuGAiG./V\$d%ŒϨ- OEG( }9Rsv EeaD)Qn첫c)bIbɳեcMm쭮lt%fḰИZWOWgj!/ +AsyEi٘uU`t!ThS%vTZ[Ż "(dJ%I1؎uBނs6AjBJ5)*PF&d6+h \ݙ IqIHVG!htPk@oJ'm4[pNj+,_xl&xtN:zb튖5˺j3JLH m ºr@ Dcka$ RED&\4gD;iש;".FQC(d9 S,,[tawT 1R(ppgL ] -2ص.n4+4TCA̬%J92jNj ]YR;"JQвw/uz6$()e4XtW.{"u]0%eL&ҭT@ŞQ,v^n!q`xKV4"2-5KMָ0Fʰ&4Q]A׊XZI <{ \ujoV1#.UUq1Q,<RTؼ*a;iF9!bENȾw c; YL|/N|o\Az|϶M߅L}6=DdX3R\8<Z8OvX=.X6P ~wR%3,:57,k*f|֚BԦdXV=ش=ɗM4X:8#TK.d?!]ۜ޽jeAsR+3zН-V?@ftHXb nZ302يւ3^F2|~)Q32KFDhij9 88kW\I2Hf~7NCb-\,rU]␈ X%TǬ<ҬdK#!6N]hE WKn$Ti_s]~B:XD;yd]pJ;6GAWg},n^GhnK?2ʏ-|?%'B^o#gG˭rđA*;|>@ 9z_#]3.ζb{n.~^m'M}5/nO/he{ c:!(<:? =yP:}.8p #:#:#:#:#:#:#:#:#:#:#:#@768 0f:PKOԁ[S@r#:#:#:#:#:#:#:#:#:#:#:dA@h&P_4Wzzk:?6 #:#:#:#:#:#:#:#:#:#:#:#u>~{:ɮz8 :{}X?Jkr{,ߝ]=F.+wklޮ6^悟]mɸr-vlwv}RWE'G9XmIĆTS>gDV6aFp?_DĢͱӈ%=K2pu<;L'n.޶=aJ7kTasSA -V+Z5؁YjeD,il0֚';eym6 =K D6=c= LW_?[i|i 7ā^=8B,l,b-S:hO6>7_п|m}Ճ_Ol8~Z]N$bQ=b$b16 hOtUL"R&k 'K6% 8Mvθ@7>/6`iьbyO|.޳bIoi%n0q]& xCZIĒ5$bZgYف;?X?YN$bnwX{gc61{v5և߼_O"l_,:XGhl$b=iogI]}dbq܎9=cE!kX"vMFM"'˄L;K6pw/|0"qL pX)'Y75IZ>8oXbL/$bֳwdiIFvv=FT0XYj'F8fk9F% ,+!L"9Cj0RlpcN$b# Bs12>HvNY-$b!IĒ JYzbkg'(=MMٙ;%?=?XA?T-iw)VyCAQ@x|y(VA CPNC %F#y3.RE.\B"8smI]q7\T=jX+ߞ:O6|]wLQdo凳2?/7'I~@;o*[ysd5zˋt?K_{ףu+RV8B>R?CW?rQ~toSl|!q lUUcz_1Yy؇$g='$gN%Go%[7vK],ůdf+CݻC$w+9Mk[LfR=ٙxh¬\ӡAu>JK̇¬|"V72e(#@%P,53fYzz+,ggXGoKS/'jpwggӗJ-JGb՛".Ё0+ xj *hÌa6gWeg2߃'"H@FQB2p%>S$ n#q̑f*|H\ cFSb{1[CƄс01z(:ks"ɬxZ` L 3L@es%"Ypc5J,o3RXCZ`2Y ;Xz ?f YJt 2B4%aPtV+o1 χNX_Lc&WXm9<e q:)l~]DSyC⣹DH~Z=6F*|Q͸eտARMƒ{nyK&d:_"]XD*镊qݏT@n} n+z}z]ztHOȎ,U㉸}Ю,]e< fQ͜K=,FRqivc0WN.8;|\.+ j:;ISQ)y56ҵƵN`uV~0w/t{;2ΡF8%8ZS+IEMr W_&IX̦?1Hdl("xʹjcD(m5Z~ns j50O8m dbH}QEN9{lpTmt5Ղ?74uÇ }wV8N~h7v 70'h ~ 81z⾋]h1no--{-EisRmyKU'D2"8Fg`LSٙ.$8)y ܵsv*`o-g 1㍔o$vFK. hr+J+Ve\U [[yoNwyU,90B(TJ0Vk9Em\clϵ)wx3u:IR~n/\KiVi&\W֧_kR#z@YF?4Ҏ_~,źC:dh$Z5!s&bbI\TygLX<GT:\ųdV³C||^԰ݍ& %>?@ ò`b4aa<;?ZHOơ:\O} GnL BfO`Cڎbxuu,|? Nrn~|L&O@S`s[-5aY{t|Xa,$=m5o Ɏ+7uJKK?`ƐoeYzJwʙVT&nhg*e,a~]s)|~~jAfqޯ'ku @,<o#_kE[ 8+iWEPM*U+:|)݈۴H1qm/e"cm[In%YS)x5-J*m4VldaO{8 ZHc p2ZIF 2ڌjR|p-F28I/!p HHG\SCLP ZR[`@la rb=ɖpr``I7y;Q4t 1F6VAaxﲃψ;xVbU8 ?, EA.-N:4;E>*ł ĕi $*4&c-55o)rZ+e+o{0D]+4}2o:˄6q+T(t1?B.2_eRiL#NXKmT w*ATjVB ?tfvRUh+Hr}ƆwwX:1(f%vFvq c9ʌB E5k|Wk^rd&M)VDa Vֱ#=h  :Z94j"kRtx0;t0ՕL:b|7@'So.Pf( FACEe °6jt*AjRvgn:]ZL?P(ajnV".y{JIa ~[fYJ[ux6%cK4Vټ2ew&a lS؃qA8g # h g|_GYۊZp#|:=G}aqQƑo+fu}2ʶ q|A\c a^ 5%DiH=zVVbA \S2du : B4(;y[wx/oV (&WǾi:v@79KJ|T`9ճ-fse-lA<]"S0Aº-3MU/z/yKENG)(HTl!S"Fؐ޹R4#yKs[ːYʼg-e[Þ5# <)qNȩbZ:^jMj%nd|\:mHOξ>M06+@xXٲ_Nu`-_~2`WROՒ?ZZ2rrgqz59Us8s.WlaFaEh7e'rjrWm~iZv=m^=8gVsCƇll90qٖsz f 4εK6β'Zob̾js&9I~/_O5ٍ49ԫIjS{k>ф鼛fM@ׁ>D.3=M&fm[vFfߥqs<8ŔbBuaAc(v9k j^ki琨I;y[i;;c j >0"1:v"x:-B~*An>2{oco*;Y]4l&25EDbD^]QY u+ռy3 ` Kp#HH E .|_e5o)jO݁ gR`8@F< I^ⶪ@V4AmEޡAiA&^(sc08:$ ӼP;(b S 95*#JL8G1fШoYۊt棛T}rXX }vYMl6 [h8JĽ1JhS6A%X)m3Q=CVBe;>8UAEYF9C8RG8aOpZ`Dia9C$(Q]Q=0ΠQ`ռOW=.7șɥdBt*T($P+(VҠvsf< uO%yKw Yo{T_]3'Ƭ6uͦ1g*YJ2WdflsI"TŶ<Mt1᪏+cظ2Ӽ 9Nȕo>@=>9$Oc7C`m4yik7 pVV(X,NFV? @ŤZmÒqajVty7&7Mt}f/wa7.rZW}I:9۷KSP.+-} ?.f6Y|0sA=)iQtpֹ6>J=G\]ʂ#/vgl)}|luJp>(֫XBH2T@1D([AB,U~2ռtEw%@Xh| ŕQ'zٍ_1e  , l6dmX:QJwDg)mqj9]yCgA#v /EdF,P:ߒ7xfkA_(uo97 (G;r+!'._s.';Ac#B{P4T(y5Il#\1;:j0[U~ot[\v[_]8npyrʍ}E2=@A璤Xy, ƿ^g'r~kئ|951}uq̒Q1뤹aK q 5O3^ڋ4$3C~PtfjHXާG@|(̷D`(1,;˯-Ur6j䟠07X UeL}[}XK%>X:n5gB= ڐhչMQ>D< r2t\x\"?W#}Kap7b7%g`2;318Ng.6X (@e*^Kn{x& "+&2V CX_;BP>I*NkCmFzy-Xx=+N 1N&y4Zb"?S䌕az LU**P lD7.s!`sY2;6a6@n"ډ-r[Oc9\>Xppg^c;;TRk܂#(k()FLT~zU~1Mp{jF9,BGkbko(sE+dR<ʻJSU>Q`1 V5!vl!# ]QpŽ'+DD=HT*K]n|\юj.Y'DՌUdyVY=jtҌ%ʸ (E|2xB6|gv}kjh\6FB"8k u֛R:ˈڞN9d~!! ޟ^CSP5@THg5EcӥAB񳘣G'r j0J]w7]: pQq=\Ix?f-|TcQ!=< N]bPpul%FxX}o)y6duӴVc:ñ~\%#cչkzӗz= [{x? _wQcV1ZB^; Di8KkeA(yE~\mQ *m[(0sRi'ʭӐ-TS<)ʉ!B~nXZi;]!ͧzx{J8y3Qet >e=P-g4K$+B̓YG}#ho(sETq?{!$"s'γ5ԋ pEp>!NB -nlhq18S"u,We+9zn]Cr p"=>쎬_fBuI }xyVgvԓ=Pg0IJl>&oz<# "Ct q@ +D$*x3-<4S=> (>O0)(fᇲ[ݶk1Ì`"Ŭ|z{m&!$Gl*7ѫp fIGíAE܇b,& a:]u\sEw{414QX]t7Etr0,H"`.ޜ6r## "F#!jCUi~Rшb5!*9!(I'vO!f jS?j4;ui|Tm[/|o'WbwlZ@9i=9k' kҎy7z :я*(E)tG٫֢Zkg:pЛ]~- +YLoҐJ䛰MxW^U$)RYj@AdV'{bW|9ڮQwb]GZiuBw_>g~x [ Y| ]Wyʸ'*إ>AZ53M$AXb5X{mYLo6SsWiA~:ǓCxi6lń~T)5cdd1j8j.lqќz7jUAE;jbƇ;M*nծ#֋=xP}=>++opR"S}Sb5|{]C5LQ7ɪt-*w:;d&t;E~kݦ8b/-WIʚ8o-47|ů6]FzX*AV6]rNpي5Y!ϙ|zyH3Iokmhxx58@TPϷ w9=B_2c@7w* WTwHM/Gap; V%YLf×Kvbna;>tDpaFtG}p^!Wb,i|B5ҥU{6KJՎw4%>:wKGïAENC5l9tG20Ɗ@urч!. @0Ekg* B+mEy?_?lݎ!B 1{;ZF՟UX|9hӄ[4;854HX{~v!߉yM/Un0ݬz#A&VR$f}{)7Tky6Zlf~9>9K[~ل`#NAChO{+t;F(zx|HY=R'Zgx{ M.|W__u?H큿 \>Լj=KMV^({sۍ_˴rb~[P`i^ Cx|ea>'wO e?~ލ|_it*U^8q[021yiq9}w g)lE2׏y;u;;V~~8 \*?ϟ/[?_RMV^L]<6WwUb~sk3 xb\tF|x c5լ7?nKem|HLr@?o+ϕ'-r6UBU+5QD)cIeLN5SJGV9kD߳~L"c'9p7|/3XqGń>[5Dl?^睸t.Ib\t1_VUZ\l."~}A:O!o>xw,Waߋaц?c-|k6ݒ+hzYJY>#^?hv_wJVEh@jtag~5yL_w~;nݧ?{O㶭",;r^/}/xxiФK$*kvӴ~CɖAs93cpzTwc15[7-Iy]ݕf{˔>7OtUY*̬Nxs;uN32;/t`w"^@Ui*%<)L˳ 0Պ7RQ+̋"0s(h$1Vz lD9z\ @V1q4ਦtpv|d<ԥVܚai?@=?m >!6雩\YF# 32W(4<-5| KQX11Ʊ(fjtیQ-,.֖?,vϪeu#t饳/\=Yg$}WUka6䕅h3&\fE_UzUʕxUc*:x=€\"XlLA1R@@ɉ/ ~x2-#Q$o mv\`[DW} z=z'E`՝4z$57$~vV'ÅoqkSߥ7p`I@$Iu4~b~RBѹGUQ_ŌX(/VۜHƌu ւ ,cv\/~ZG9YC"b1Xs$C4$ "Qf)2.%gTr$<[ zmrF{9jA Q8Y`Ғ7|?6hGp2·-*_joj]>E&"Ha!4iKiq I=L|Ts0Es&)  djs>Zc'YLnJ08#`;κnIMOk 5cG"jPfyO;"K0 U$3S 3SX^TcoP R/QgKJ Cdmi nYC7:00V1(TL\;܏k']0Z18}Y@F/%$+VOG'}qAդQe` VN xZ#8+9V0binAK %іg`3 04f;>.𳴺daMz73V+|@v{O;9 2F3K"kPMDL\T9Np"v]"K 0>Mc EQ33&NdA7Ȳ\$La,D*b/Gs^gx"`,GiQy"Pe6ss^|펋r(yPyNxqȅ?&H3Cba a%@)IQ|mpyA@(]_:"ek=0m~E—cU5&uǺG~]$ǻ3mk?7zRe O31V*C: # mE+fb5DczcXZ禠>{|삍qv"2(ۉwhqZ(⨤OSL 3G/= QOQ^ռ ڻf4FAAYѯFSɒk:ILeQ|_ 3b3 9$D!MJO'8 i1t[)Qt$ ?|}p6Yw2f sUCpe]gl ys*ITl~L\4kv|SVIX̗ ,Ly*qU nf:InܴK7{nah|2LfrA2-9R[*&#Х|^s.}/+x/Eۧm^x}xDp8԰⫲E]eMi]ePe5())'RBH.]$u~1oP.aaŶ5ӕ$!yd#͈V4v >ٷ.2[4.%Y,2[;_pu>QWo>|UI?ާ2˓ |aPB.}Ϙ~lB, &an%#pZ2xͺ1qCO7to.%ozs)XΦx(~73 j^)SlvvVs̋L0MV qm$>PIW [ujn٥lRiRп*6̳ɾ-Ͳ|9f\-j,?x5tT{b8zTo󅦦(-ܷͬôֻ"&̴Ót,7i|w-"b>AWG~.]Fy<3vfuOV"f2S%} ȯ)^/3LyŘ0)f4"Q:y*H#I8N(I)(ˣ}ɼNM޶B]:b_ew ,¬ ~OfD曲z0d;4_i ! >YP;;>"Ȝ7Y6_~E=e}͛ffp@vC^q3Y ^ [8ZX\. )?Iz}orVa=ltf|uzg̔ù'vzͦ5Ú~ò<ﻏǷr-fs~ӇD\g{25OSqﲿ 3LkOf0Oa`ݮLΉz$DnVp a\77a) o~Piu"x zdyncY EzgT@:`X{0nZ/Nw>=;y:]sL~taՃzY"9%Ewub^m,8Wd7^nx#evxT2fb8/wX/D4,ٵ)RLQc&223IɜBVfalKt |8VPFk1pjccEpn &"\ҔhQu΢Dt;#8 xL Tq{dnrx;}O[ ?j"7dj폅guf>:־maԃ߮5O:iYQ;_l܇Ɩi]w&vnXò+g3Vկ~}O~#T<EvYp {jBl%Nu{i Ix'<ym(TH5MB-14p_jtapt>CD%"V{cX\IYiP\^4sjH4hMOh|pTl]dOX|ѭ[!DQ؟WXsq h!t{ xb㱇TŇ->J-1CBob2[F0 |!{iÐ'8Ǚi6V%=h;58Z1Ib* 6&Zv &T!=k1ۣfUVZ'$$MА8%U.B/ap1OYխZ"khMWwOc" ;D}_TFXs`;;Y0A.tBXˆ<v p3K8KJ˼++( 8"iH=v 3yHа%q(cCLl%$|z +c4F10s x N`MUìC`l0|һ᝝$ /|=d$u䯣.JIMiUˆCqfp cWF[GVf"Qf)2.KfTr$N.ڥƻi-`cJPT,!pr%v=qbPDɧ%ⴢx;N hIT>AّT%<^>`6ؚxC-l(MIw/߾ ۽ [N@W 8Uwg.ەߘN2睌jz]!*g{]1k̾ U^ SNn h0{hBE[SWPi*ej4;fq~e-wjOCU[}z ͞+&}`:bqFD^y%a!`n)?qUMq|xK0WYOc _>ǐД8!* +Q|mĩ߷[ʸf}`N-3^l0IydRq$H2S#A{qUFU` $̈m#M sdӲPObzϣ2I{8 EF-ƵeHR:%!:Gnjۆc~/]J Og|tI =q`4FcoE>[$as`]iyAWȽI&8MAȾwQvg&uHW85*[=ko+E_v(6 &-.Xil}G%[#͗y8ȼ<Մ ķmU.smf3U0@ɰd03FM洒!`C٭fGYf,= D~&Qt/V=nyCy }uv%Z, Ҥw/;8:c gXffKIQB\8`"6\ P }cgm60ngDfYGc(1 2$E[OՔ [Ssc ,Ȝe#57&:m_E-)`'¥J1յyQ@CRO++p3zG ,H1Jm0]y2AW )1Z<;yf~?ˑ['K%fqaUSa?1y/jt?ATZTzE^{9ʌݳ6#r D,~awIg5+x+a**!Q1 Cߐ%o=_RǏ1 E(U=ʑ6 &wuj-U.g8ff$(Ӑoᒮ 3g]-+YF)Z0EW)~$zuqj?j-XqS1NjDUE,:]'.)Yq'}=mb&̮#XVF1rwͩo=0q-鏲Cx١M_* /wA"r}RQ/L*h^'ǣTc]7ms]sX1 E.FMx)^]Z f gDtB9Md+WGv9>0ָ`Nf[*Y=\G/ōC!KoE ewWC//T TJfůmKqfGWi0W%ZKx+7*l"] WˊO^ErZ0KR"ZJUA;`H1Ř(\8! ȪGZD1b%LMcZVx/zuVǏ1# kۥ*HYAoJSKTxPrQ%5`'+?nfT=bjrU#*%W5H^ӘD zVOmN^]oc2]4tmKx'E'ńO2˅ G%Xo7KYHőo$f,iNqX[hL`UEo*Z5#`Ž2O L#yG?G{Ě"*4-զ]scsG0?f̼s_1%ЏVpaTh镄vi;uK'9dKwS])|x}u(W\=zG" 'faڷ~Ⱦ_d|B`j{;NFA =Ĝ; ;RI~c>[=`Z[U+l-6HSK\-DlI])+cFxA29wGT3J(>]2"ӄaG $9MC[ d?m.Q5/͋JKa?N$Gd^dGû;ҕ3ߑ<# ܐE/hvrUwvFtq10 "r)g|s#xVX7/$aV  q;J0h7ա9 tv:]PԒ(h n7 }s#B>);Koc=6;;:ӊ#%HqX9_KmUH`X"Xy DޣaqڴHIGt|>FXd JE]T"=q%'V!MA;QT6y*4>A;$Kj)R1tӢxHeaJ-<f]~euJj2GB(qC(w 8{fNA8kMG~SOz)?EԫHPZ)I@A0{*ȺPHj9 29bԧ] FƖ8WyśN(uF5tF`yJ*dcPIU>5 rwHo 2c ޚ) G dƍI`ͧMv҆ Hqz)l Qa9Kbɵ`!8Eԩ˨jK9!̎EP/;v^ g*{`fƟ!IX|C@=-gՀrk4F)bVLԠ t(x~v7C2kE`.]yep%,\pZI&gFHs?!e%{A?4 lI5.f/1?5i)fF^E4f<X1ޕ" „OFS + ֆ_RgC1E9hFmCϋYpƷ`|M>U)6xJuNULZAq&+G*(dSݬWo  x*npW !Ԫ63hEQ%XL_yxJxLG"/B%&Ng!o2LW L GdSqʁl |{dFO,ƇVt猭+Rrh.gk-a2a%aS ẗj[cQ cpZ7 ƆlXEte5zqfL!b) VtwnwTQX(G=l`E`{U5y՜FfDuAn$Ko͌QSD/:̯׈ pIn?>T?5P*3dWоxithI{d[(2 B)rK(&gMoG`MW)~(6&Xb]Y%8œT0ܯB{MVi YM<ynf<cV9Ι(mF^Ѐ;YBLZ^WgfȎO玮,WMe%V7!.[;ֈJDˌC40v @^a#ԟB#4KjcŵQƲB}ͨՐb˪Ff$ywJ$F 2-+ wV#8CR0嘅PQ No7s8Ff$^>vۛd*3s1#!m9Q\Q49NLH`ꁑ94==5z(ŶFf 5ͲPpuDSaC(tRdƸRNC9*O skA!4`V!}pxjv!ȉ[?S)Dޯz倞斡e??#+ߡ)߄#~X/w?Mfn'Is匿Kq|Y~ie9w/]}9m$8@8_3!_>_g݇g/`g+UGoqzSL%v~{-jÿh݇ &B N~9Yiw}|U4ԓ7j:jo|5R?dm߯MB~O꯿m|Y/j O&lݷY;'+;sw"+@v>QGI -A݌l \GO?3 P)DgWmɄ B4C ;Hɣm '??_'^нW_C?!Oa! thM%KS{h!jɜ6Ϊ+(5 !ƺ"l& _< >'2qs}б |À W_椽{@L볁~zټ|)PRL{ޅLᴂ^N7?7GgR_ Wîq=&^8ή)ƮؕQIE_՟ɿ~o>n.ƤHam/}1][7+Bc Kn2 becO@Y`m;VRT%Z5Rߔo^ک>$S-* tC*&w703mF2KcZHN5:Dv 42 g" ,` ұΗ!ǨE`oDXFa l6R5>w8lsy rOA" +D+:<*u.$RkC'@%@n:z澳x\/?A@V\[,?~Tߥvx,#QLdb,vɋW)ο1{a{pqw\$ @{c:?bO?٥B0n U٪_;dC|qթxMͨ:zWh^>s||<p_9X _:+^~e]iF ieB4M*^0݃w.Ƃ0`~>o"BKU>}TVށﺡu4{JHY\% I!rK9 >8h씭8&B<8LGlrxۛUa3&E*a!7&,Y`{@/cŬ|⩸-&uvXhtI9+Mnr9ǀ7^lk.Yϟ%}')v|qW$I8['4 P߽~)%iAg\t1nXt1G; ȇa6Bh'~𿓏 yEϙ̼*̹ +RcGrb3JLuR?'~`b&9H,s{r7`4b0o%呅TCմ>8apbGEjf[;/"8Ϯ_;R\mbW*Y6)@X/=5f7I6V:h&> 39u$9໐͸)^7#] o_,öв| t^Ǯ۟]<6̀qWjL۱w]Šc/se>Ud }NGb/2KfIhڟTG_;zy/G|`Տ:zp֛)_ôs p8,8TlYol 5-QW\qoji7q~Ӱ#MMmrs@r6Ŗlcp &W kfWWn2|OREJ/Kq2 0w^l&1%֞gH>P6xz`P,|遁)Tۼ7M>._ܷ+WcfaH]q C).aEq{.\DAoKUn > UWYրb=dP;0M/O ӣ`*U<U*>TV* ZEZ'AZ2|<FV1u+ K瀵l,I2&2oTU2&3>DyH1}``|moR϶m!Bv˷֓تЮR-; SPr4W롸 -{/K98& 1cǹ.a\r.23bo@9J*rڢQ]C;s𶑝LI6϶%~p4Cxd~puo=Y11:9$ 2Oi$x;~^ )*xI[Ѧ:N)xh)ꊹ%;sRYqC~;xC6A}|OB6u=Mjp(xu-$|0aVE ia k;kz^E}^eon!I Src@+RmWD+h{E"^mW7po07#^_EA?7%5Lny%Cw.~sCTrbvtÎ0E#J},]m?S2H5T^LX)zw)ĨEA/V࿧?!Nu7Jj˅/.klMIH@V,[XnQ8G^]la2hM*芧]Vn_X>6IgwbZg"!8+ͥ?wo_thtJ.(3 `vnUYV0}7SGgN~y8 ,]hw?,K;H>-J"xo0dҴ}6mth% '&~ϰAq9nmi[lP1D[`ڂ-0m LӶ-0m!!J&mGf-jf-jLCeM۬E]mY۬m6o%ԣT=\w!Ou'r)ç\p݉JY4 zzު'P-Ҧ"""""""'KRA9Opa֞~7;=͐ ,';3NY]{+7FPwTO6?Эs)IB<2`d LZ*G=Cgu)(by~nR!r"$b-Xk-d/a\wxj6c/:6z8tGBZ%Ӵo mǽBh c}9cQAxWNKNm0skjH!) Iq0ǝvk_/bV*m'%Zjrmc6ֺ.~ tf j @Bp2Ͷ /kaƨ!OQ3vR2\81C #0㘅Ľ2?:;k"AJWSʳp  d4)ik0ddLԫc-۱m QwriW`!Fg19Jt.@^ѹqRy|"MH9aX!YqɱBroӐDׇcY8dGK-UgS z6<@TW VPΪ0lل+J\Upt鄫ư3 W?Uj~. ̓J6\=pŴKtu3 \3_;L*xUP/X sg8'y/"Nמo_epd#1Q0MAC<3_*fH]>w_>oWtV^Lc'%d7p[o7N(YURV(y~s-7`W*/5-#J|vA:cfv|o$iK @Cͬ >D3ӏ~]Φ͓&#\>y:5EA$X'sC57!2gB1Q:R0(wBPca z\g FHӝy`܀bTQ.7N:GH"(&-gĘ0\lg ̻֒\ `9%5gijo0B )&<5^Sqc AxyL R`&#`aJ3OÞCf$CVp4ƙrg3d̯ݨ;yVgÊlaѦ$ iyAI<)K#C V|R0Hm kpjG $67!N\S 20WNeagh FH1:5cr>J 1+7qjo0B%,uV2gclAo p&(hN3 9J FHvJ+ 0w% zc X Se F:(3$$)9d:% x?gfCxk̞=/onc&7IpFHeShHT$\f(,'-(|0T|-&܍KV,`6B4-b#dTбɩ&Teoܸ )I49f;B R:JFK,v $I`YZLE"ob7 jx9Q#51RP!<4C'THذLek6B(V[gIJSc2L9<8FuP!<3!Mx94Ө ! iC(^^4H$ej;Xr!%C(^J;Xly+#_TheTJ˴a5P!,v^SQgVLdu!G&9!X?-s 5AN c=e1B3 L@ZED3[!b/ mN:N.E) n+9*Kj|>BU[.ϲipe}b8df!t>B ~ʑ+%vI](NEY05q5Y]k, HD%,uC(4a$}7 &Jvu_Lxno0` 6X{Wat sNoE+֧樞a-{WNT_sKK+e6 \q֧na{W]ݛ ɋ<\R_\$O/g.M/g-}=n0'z;pieaUoʳKzW[[ϗ\\IqrP@c:aFd?ՁܶiENj\zL;>ho Wa%{W]W-~}\A",.#zW.%/pÕR\AHާd3W]-W/< EbL$Z?ƽ+.}+ϭ/< g:Gp bUno p:7Ȍ0;Cl4W3˳LrìbԤ%9;6I=WJ|~ S%L c4DX"2fYfSLx;RDdL6MeYnLl;Nnn&rJ]L hgoa2cDzBr?laR~?ߺ^,|8([yiՍ]ل)6z.OBUwd_XdSFG{aE7o%+cw6 Uv>į+p/ӣÚV璈g7'(ʅP ͖K: XbY{70]S׊nxk,g] G/nqUh~m*T ?]'-4b+kc!(%;WJtkOǛ!&(|H s30 w kR$hKB\fMTJjbXf^w~>8)S*Ӭr lt8$0Iے# 4NRii% :[ԡxKlaT~T70X]v!Đ=;M2%g*`i G{؈r'T@l\0lbh>ʌd ehNU 6՝<uGSDh34I6yf ccIXїL[QX@G*:Zz%]kR;[SYHrfYsnk*8#AetA~[PQ*xesukp5bXC+/g+Us&X683ir?CMNsW!lȌנ/ulz~j>)ؑ?d᝛ NWrG7 &yŁJ Eg" ~pIJsG'p M|efnG ag:D._f~1cs .~ ח's0@ x|[0咑f/kf֮`mےKUnPCᖥ]&\,nױ4N3_aâZ$J+⩭j@txU"soTJ9Yl=A&tD..=V ##&E.AD<ގ$?hfI-m>t=O&NjܸY":LlAQ{&}s W^zK R e.I= Rԧ7q\qy1> nA"X.WJx%8=UrJ(a 3u* ,z-Yfou-|ZcԜI,>:n7,R*k5r駳"1LeHQ|ϊb|$4s띛!R|g6a݀k?f7")N3Bbjkw]4I(\|o7`UP"Zduo,3"4g7mq*29mcV3?;އngA3BNleVMbFȎK-T}#mo IAQ 0sV)[k#쿁D*}+8Gv?z=י GȌF{s P.c;~-|_E#~ֲύ2iA0<+*ϯ+܍oߛ@)]jd:|1F̈r:Ɲ#Bvvi7K{_Hӈcif&y*p,89hBx.[嶅uSHLųaj,* ʷOƨrv ?|]>i ZRx%!1ȗ|v6h6[2S; {הW>XXOFRt$Er@ ߒnm"/#.Uٵʵn6fM 87jAꐉ<7z9;u7_}Od|";D䧻^pgS.yHhfi52_斝]V2~8\^d,{LYƗ`+}srL{#-ɯk g>\&*>shN)8s" @@^Lcz_ ~K# 'Ksuvqo@Zk9HUZ|.=^m*j&Vm>9l0\_;/fqP0ò]T hMD|2hxhH'Tܺ~p15}tG*qk9<`>cX7t!?;z,;#䐻3`?Gc`:wV"Ff0j  is*Mn`U]e^-C l=.5<=J\>cB6KF'Lwc籓 }^ͱ0tyY!U}I3c}RE'v pl/E:?-Fp}Du,O2"bMzzYIBd%/d=3B3^IfȺSY33 g e7ڨ-BտGI=t!m^wŸE$Ò!s6A{xc|]2ݫ0d3F_"N}&\A0g \tX 7$QMʼyӋjf`K`K'J^ɺ" v=\.TwѻuovqO_ءFKӄRϠf[S1NoiZ3)f1[{.lLw3A۝{x۝$.;:} +͂l?JuqzN_bX"7ڳ_}]0j Y~5" lܵē7`1)[y1gg?(Ca>fr"9у_MÃ0qrڈsE7Siqph)I-ck%NWv*Q9? R՜ON?L4@9ltks{-?>]Ѭx#/ޕqd/b@ ,e%R!)Ǟ jj.Z+"ݬzU"=Ú{g1Kjz 1!?}q? )C(2or rƜ3BHjUr€C4r Z_ܴ-zIcpKěrՃS%&T#C+/WaYN9Qh9NJ'+k;Hr!>3#4:RzPڰ`H uBG.5M.AbB[T9Ux:QI5#N w,+S"ɓ3r*9ܹ>uD\`qiup&юʇ8)oojb*U:ؾN0KԧVD[~_:fA仳,[oDsg /֟D$أE7`V'@z4 ; Yf3ߢn&HD)ߐ{3,{!(ƽ/SF|taP;)ffS r3vϊ@b>YV SoddzW`a6%}V]FGL\/5dhJ-P1KA\07[΍ ca:a&6_>ɼ~ v:LAߋ]zQBツE߫`Wg̬< Zn)֋S2(rˊ sXT'ec"i׏PVмC9\](9L0+ӛr3mͼK"01)E 0;1hK1/4p'ıq,:slG5R+*)dFR_im*Kؔ&^n3_߷/z!"!iMG41˛f]4;QcF4B]1ޚ. /Lj73w1o_>߄sTV&4;]\XY)W{>(O*Bad XUUen:<) g"D % tyDyWjO}d`-]eob>4:m4_|Ο.Fp}ycF\&?~P,Nճi5d>:f0(% #Gȡ}S[(L+c]at96]0U iθ++%qAcr !X\bc,`+mku];:?+iuq#VƟu\u [7nݰ5 $0J !snNbH\C5wwNLneIB^({M<t!edR3KjCY(HW3'x3Ӧ].r4O?~' dwl:ֿ"Q\&)ImRwF'Z4yHUY>0gxVﯯ$s.zIy{<ܘWSԮN"2EYL)Nc5U.0 c.0 p1a( Z4Ahmbi0l,5;1¦}8kGi];\K"Um6 U-o8i/`̕W~)$eD}X#0'dKuBR&uKf0ʛ+2KݢNNFwWlǃSW]3}H4shw\F7!u͛'nm>8'M S9oKd%BiK82bbK1StHqooQp>5QE Ȣ˷ lkw+~q Ԉ!o_=Ykx| b@%e?!_^[P۫+`ŋ'}tIa$0v ;l# Ԏ8.U :l04BR1$pwuSkOfh:Z{fiZfsHTbKHɂ #V[ ,fijj!RECUov`+b>rKdme?X{hU>V[`I=%q:KbnO@A6]=t޵-mz)*GSѶNW+"m4"# $>GQW#{%G gW‰i2{#Xv'Nw@nPB6Ppah&W>I/|خp@ΗaD;1X&8{eJQw=5;թK\}TsDj7U pVtSz4LӥqFq$7sBhܱbG*&~?ٷgpygQSۆ%mKazm[g=Ad $T# hNK6` e'^N9zG)Q KU˝ 43%(%Ή^ ^q bF1vuZ 4Hڃ7ŕmAL*㍉%7Yp-`v;U XZ nX>,_ˮգ;a}mʁ2^9P+x@ټr p *$HX_  57:.D4ME7p4B \ :g8eUeWmр]y9W\6Eޱ%b9fVXaEXp":\j޸ony"l7h@/!J`dwqM<u0r ;OaG=NF܇QX8Ep=5RwP3}4wIXutI3XrLF:Q_zjfc}bX?IU ̀k&x6_M, + #%.~u> f{;ܢlN[ -QRKbd# fYb~U,V!P7H)ʵ}:{|* vOѽ?&TIG<ֱ Ek FDkBmek]` *qk fg~JXH@9~][טrL2k*x#p[F )9wEf%0Fֶ_>L7.I\ڜNmiQ] $jةZEi mވtJ癠G. 5(%1R_+j-FK?#l]X ᔵofVV]8,އYbD)[w#tu^rc(QV*0ᜪ D3_$U 5h:sJ556`ƭ^ NU Cs^ T%ǸmuebO37꾲~Bl{99Hg?q we[LqtIpfXoةUӋ=-Z|V*o?;>8p2sPJXƮg PodGUR (~$@ gMKo²ED;cLQY. FFY,dۑ.$@򒮻>\HvVMK!uUSX"Ag"t*ޅ 9̀)vʯkƏR6=:^| NϦc']'o3ZOs B>>RDU']OjϦ4.@(A.JedKF b@IL L)3Y#`gXk#k $Kzx3AZ05FukxUδ ⏍h l(DJn 1d%$`rE֗VSCe[N&P-5[XQdi!W(#s @]vL^u]‘ =A^T۞~.f~`MbUYԹ;18t![qR/)*0Q@]u-'X"pnǴSeZBʹ"+Ѻ7Vi[!ZZS`?8%u$FZ3N|FT`2]7ҶFkI"גj>&v.x%.%?.1$c._YHgțO8V޲e ] $\!r2H>/S:[:$~b1-7.EQd2%1GPgc3p H [{|h V1o,͗%+2 2Уh@ʞ}YhӞl%ۋ?z1^ri=w5x}(yVry8H=!'!p:=ZMߧx<*]ZO'?6W(MM GQqP. P4|nm>xN^y5E`v,̡f>5~ҕ3M%kjUAMX]VFj/d͓Owۇl\nyf߬n'7ԥJgHɨ R6ފ[)⠼Toq~&#cv-Xy,s̋v F⍽R0"ϩ 9\wbp+``Ū|{[O~{vbWbȹo#f.#elXvX_fxESxGw w ns31~yw6;yD&3oJ Qs}_nﳵE^#ZC#%VZ=(|-z?mK-+ ZX} b Bum_?׍I \o͜^B]|.]{%OO@/NyW{($C᳹COc3,/*ǫ/֔Lp_45`0[rj`%Jp&bYԙ%դXKs=R!%I8}VZɀ"ZN˘"[a3@dk9uxS+]ߖ,R{ CU-eqP8=P"ݜWgwڠG*$ig1\2^&JxJq%}6Db>ì"R7Hv$wژ5r\"q fɘ<9/,Ĭ8i8KeuhY^Be fE '?}a?_zy K_z`sCLIgK$Dn!Sz:{ul$+9}~޴̜$;'m9"8̧oo<|=fZIr+t[~|ȥԡs3 *4uV2Ŭ~P a 0Y=}~l EN rppڗb{./eBOï5v԰]E)d-IkMro/{Wٴvle?zNYdQF{,~ #JqG#\ޠHThȗ޹͏* le,IiOtouz'mF&*/^J4ۦ㤽vHTFB&2B9(I퀽S8"ڮVf6&1Wg`+D[zi{Noay bAL:\;qaW/7 ATooIbb*s*\?|Oi(!l5%)B05Pg հ.(3mBX[0.WؑL(g8eiԇ9aa%]:4+[l꯳D`T^OWqU8]F)2wNBDBř6%J-roJ"{pu93D;XD9G3 LFlψC킬d+=kVg%O9ĿA O^>]0kW6

K!KһVSă"8p pa#ao$gPoA]#^9I~MD (\ $5mo9$z_\Lp$KO|V"b(WTB:c9ud_L]Bb~o'jZ8{'IL(QդlcAY9՛<e9SdNVGظ8oJT6I) 5ẁV8W%m]r)@J=|^*o$է1KԒ9E0PCkқN`:O"_rJ˕i$ "G}>!EaHH$p+G%t8sF\2ZoR~bo˜tł(ɋPoZVLJ7ξN_2yMVp hnS*-nŶV{ꛂn1o_JPp2 jZQc,PEXXU*7iedx7, N So%V֋K%'wb3F5d j+f53%"jfQy 7tgjv')PIjp2J@分xDnyǥGJȓ6Q6+!Ҡ˘/yxIՍoֳ@{e<^G&R SY-Y[-/y]@1o޶:d{5,EIu;~.ɣdA y.7Lʄw:]Z!*fO=2%Q-︬l8C☁d<E-護r 7KG+ X+܉m.T]/XZ3@f1%(-92p{3~vW_S)m܇X=~[gz淢Dӕ:S2AksZf?A {ȕ0h?%{$3j")x(;bb>} ϖwEϽdlx7 1\V`Cʻ Ė"v?'eLv/*};pIYsPG4jX! Q1U*Ɣ:%Ֆw, eċvC|=%9w̥*?#k.7z;*e⨬&t6۫Lh m2xұ AQm@4^qVSl^i׎R+d\T#S~P!,E%%ͭ2I Gd8BE &xTc@ UϳGz"&ƈQn3NJҦ7aT.>۠|\JH<`$Z1.UE-_4weݙ9fLrfp2f~Y/c("~5*KY^A1CL>7e%1 L_NWs7p sQXԿ46= ~yis Ds.,Ai5̧O6)ii8J*~xuT/2@cD%s! pT0yNS(=䤕ZUӣ$Ж+7IW ޓ6dW9vˮh`$II>-mٺ,ѶZA{udRQUEydtf +4Ivr(+ V2,^w?P/ ɹC L,CI/.Q  f@Bsx]?{a`ǪʛOLg|Pۡy1-Q3`a {?5>r3d8>yyWs_8\9"tTUjUI9I  x菋r;8w_ DG{E2a4(C9ի)U@R(hr1х7[* @L:7X` #ZV?ٕc\^/U814=Z]nP9r}RkWY8:k<&Hb!(FZcpڧ1;g'j,Z(&7 [xZlԟ&3ǩYyԕSDW`s!lv?~}K -]8_; \m v;Iš3S/ew,aW;D͞5Vr%5eϿmeRj46aH}W>e}Zy.EjrEJDmu/Q68F=L7zU `|m1K՜q]+KCiH! װ WU"ON\e.9〆ⶱ$D=,$%*L@p:CJ](جvUrñD{GJ窱՘H^RYWNXG{~8MR FAF2`!b~(̚CvM'O9>Pj!:ڳOáarDTd`=NuY jOʷT3v.AdJ?-WMΞA)8_sƵD}>/ܑ$(kRriWp?ŶN'Kw;lNWrܸ`2Ii JkSaNEq}cܧSaNWV#6G5 Px ˩IR&|`k,$ҾJe`ŧnpqC%^/Aq[t c$8kθf9*i,m&DV7J 7 ״ц1>*7p"K}oC A.GC]~D%pb!k˗:K'3'Wf{(+?ZME:'c愇dKq*;8U2ze:BSYGMT 9:bw&fiE,{9||lvb3~aXj+T#pW!"kFjJisO~0ng$kL+ :Vr~WĦ^)h5(- #G]):E郦.z C[m"qhx'~iy@ѷ y|O=S8&%32M"5g^%A

<*!iwz''zWt v?%%SCnMo/Y#XbԨ_EE1,r1wQ%Ib7-mռX&>k!nxX;xJ0{i*DKnRY)@読=5=ء5}׫  ]N3 HFc`A{6F z(-x I 3 o ߋ{Dt\(%u.8Иpٵ.$DV7#xmly,ce3vS~}/e+i%=ڛfAuwoKtVxc(!O5^!35j Զ_~%x+sy"=lHl{xPCHRryBhG:MrGngF+?&vFk%v!uhmэlFh.]o27WllVfb%礔&by5 Km_c7YR23Z-}uַMY(5z"d"s}ik_d Vž  $rF P`|t `$ lZLsle@ :-@1̞tKw߭Ĕ@qEB9u`oA׹ݵ47Xt#FIb&DT@^HǃrRZ Qo8 es4^wZPwu;iʥ'Rޒ{qs  5"Uy1O~0ky9HHQ'A1[xPX> 2exQ >/ʍ FgwFǖ Ցb'O^"e|16 )rq_huM*T 0l\ܗpF\)MgyllA $MHg=NW .oL@42.*F d}WJ5zȗKpIC~oy2Yr98D"cSrF&IP IXqsJN~X ԛڪ)Q~ql<_u?lF3%^^PA4Okc(CϿ,]x7^y4-KᬌYPhkno5wnjtjV1pgH%݂O D ݭ .DjQv!J[1mݲcek*JLZ}a1-g3fx$XkJ42VȦ9 U,i-Њ`ٚ9{vN3nR*[qr>u.Im=Uџ&vw$FE:^.H9:QŢkw9dMs%DzGid '/iNr4?VC*G7\) /l+Srw}ZddrDjz*((KO)A:/jtuƷ6gd]9Ss}%x0۹::% sʺdmvcp[Y$-Zʉj<%Y,˰&tNKf/gc^yۭz\Q~'DWBr: [9f8ƄQ2,B(CN8vnF{s!LF͉M@ De؂1ZfINkb(oINH> MB yq`JhG5pa mвkb1# HO׀ !ZsĺnHȻ|Z.ǫkǫ[\rhO9pg ["sBhE%fwhENVLkۨWܳ|PH鄴 JΔSSP5SL┝hGz-A|>輷19Si]mPE8:(cwf|ӱ>5Bc| DGM$4hjcI0_+۪Ϲ005mGiމsqe{ tm7Nhb[KpMfOuY?.LJ@㾼j04~u (Ǻ5TeZۓHnH w,rL^3L,)43]z 2KY*[! Fubf(ٺ?`.bڨJ5\ O7UGu]Wv(#[ǡ#)(a\^Ʊ'@ EΌnDo\DQ[$I0ЌsqXn)w`$f gMKJ VqnQD5'Zc}FئY7?@f)xnn7$b9&_D6n٤( lt37`2ozcx@o=_~CQL=hh0.[b&9gwn O{i~WvmMaKf?Ex[>؟W FGfZ=^&ش#VPq{6R41FC?I cؘ8ѼnxD Q2&.aZ}.&9 LxtKDي/&&7珐0{Nr0}qs|n0aJ#v>NߏI˗kx`î^Wx}`_/+Z\h<%ج1AtFؓVIx}pg+huwVj)ܺR?om'Wx{olP |R o zRSJH?bןnF?y+4~5 ~h7? o7~TKo&}P!mtvO8*+FnHQH\JX%2j(H0= +E-rg omh#.!ъJR;rLG)ؾ8n"R]^ZO V Ī|dj= KW2ZZ+pW$jS s_.g|4&2R7DZf 4cJFa*YaWLn۬n0Q"z?@ C޻K䌖&y i[=TrK||de0hZG`mi`20' _)VX|~>H\(`pTMP/DcRTsֆ&oᤗ͑2"WBkOrP xOS<0 ٳ7r{If ^HIsTs#5s.DyO*bpwC0֑ڕ6MO%p}>/i4J1DM)4V)Q;L{R'ZMz:ߛi~@ըSkn}UhT6i S\yEIbeS AnC7t}Zljm)nM[ f}nK4mvş;0` u뀾Cyk+GpWyA&9k[(IV挽 >3˰@L/c2MR'OԿ@238Q1F4FJt7/*emz)4?{WF C/;plwb"fv;yU.E$uX7xREr#,Q2@"O|X0o:nύ}3XdҤb} 3(oqbCgz}ߎ.{ p-F3}?9CӏРww#-,qaХV,.MkAQH.!w^M1<.XCiRz`)]w2.f΃ާO5^ CQQxPԧц1Q;+~sgo+H4b`4H=H):Gǐ!Ƣ@1æ0h*hi[6[Z(6݌g^bK*l)%A8N@Y9Q$Jd=0TX0rfX\I)p?[S dγ bٳh@)`09CFM% &g|-64p0 ..D:pȪ"┡X-ZoJc-8(o{g՗3cLjٻqAc69LDSJu&aDHݎ#Rk'ipm_ﷷί?~/?u3P>~075YflW,k j.d&VD:j"uw~!˗Ax;w&޼.,EnB|;]$E]$E7 Hs+JR0A,dM(!V)uV`avbwR lӔb|7pSix3$dчxKj,ɟ=7Dqɛslޖg̜1shtc(d #*Gaiތhg4붰( @>4T9kAkBYr D)aC` 0hPVhWBo$ԈT д3Y;{Ge,5 oi9>ix@nGA16%OME.W>ĺձ}y1E*H`\ V1c٢eXEb,}(e\0eAA{%,,oJߔ9%b}P+ԉxFGJ8ת߀&Ô(:ZW=<j؟z,! Ʒס_M݉m 489t|RSv&ϼuYgAxC)L<]PL"(Nyc.|rt*zۇ^w(N@2cpS6\Y`b8JS\OS%/nRs@7jvb5atPIQBij ;4+^ZERH cCD1GUPUИgTRnRs/TJqxtyq'*w* A^op֘ĥf< {q䭹¥<yщ xЉ xQF'#oFk:J9nL`r(4) SxCi~ݔPȻ)n4;4HIS@Y5'4NDa꫑"O[ qXZBPNsif4,Tk>JO m@W߯*VV-\^7Ł*Y<%0CH0C(ÐWDxA,R`DL} QQ 9UzɌÄl۔P#۔˅zt 4+I528&P{8cdݸFVVm7Wvk(Gշ/tuL";"ZW[MkًpUz}R+-[dH9UMt"Xiӏ߽3n5‡M>ςYC෿?OO>ka*aY8vak04T[_'- g4LODM㰗~eqAwt:-H@\6`hbHXWZgE 얐[:*KSr㕵c|oqe m6E7[pmiussp 9QZG#1" *21AX *zRdTBpypZJH7E( GYӜМ=%2q a)^+r`c^ ICI lO=[TpAYga(|7aΒg^ƇW1ߥ7=s[՚:-߶7>5PkJ"u,%ae݊m[{OSƪͬh#- 4ڙ@D~*)^rrrHUv`Ǵs|pnnghMP+wdOIt_>4RrB?NV8}qyӁpR&?ˠ-92M6:)SAlƗrNL Ѽ[bE"/q26~eHcw#2E >lfVKe1ޣkoV.+3O$/&؈ˌtCd8jp5]# -E, VFiDjH lTfϪjS||LZd8 D^3:RTAc-T&0\vG;NAXѸmX!u^s͸+>_)'C baТ" 78Aj3s ]I(㓘q:/VnsC35=3\ո5vkNAug;OA4׷<*Mw4@  윍N5s` :dK?͸7ʴYY2oSs^.o9[mM:[i4B Y)ST3TA L!\.u'Ռm' wެU.6n(Sßm/pᄈonouzh51@p"6u&$NyPe(y6q94+¸&e eHi>{u{{=aB.?fcm7RA(GRdgԼ|X$Yz(u{#f#q g*h D3ĝJ. _5@3xpY$R"8-Vh6Z FBXt!= ۍo݈K+&Mϭ ||xs",lQoPuԓ^@lPO1M^w+9ﯝ3FίFW8m4E7E. _RJ-bݗi_dlƶ"꒷puXo9.TZ15hZ;* o(Jrv.N+O,nZ犺n9|ȒN'%UĔeuteN *o?=wɩqrl3\4 I1$RZf5?ZGDj~<ړ#_ K6{D֐-j ;)I؄=cKHVRijrIcʕy0nmO@77IMq7hR2+oi3DN3Y,`}|6G|HhTx9Ʒ'4π!GH;3xx?Y 4N@Ϋ!e]mNrLfe'.YWL(i/bs{Q˓u#2 _S|̇-r*hɫ>sӧj@s#+hvm$%xBaý.؄B#SPw[UDݺQwwc]´$HKJwSj0 /IyNd_U8U2Z. `%v1R&>r4|Y'p&:08^j0C X` g u~͇?iFQ`F ?_:w:? !yygY;T;SKOsʁ1w6SL%xmƷpVC1pu?kmV\K?&A=N`@bYR$9V<$Y%*,SKKCWuhQ_4:9(hECxhEsS Eg372N``gpg(MƋ[lQՓiu^(oOwR)ܛ97٪g'\n?2F}X>Ɠ(oq{-8 J@0VF_Z0e>1 =2M$ LjeJZM8# fRьAMJӋ@CXۇ)m#, _CK>BR`|B:TFXnPAa岠E+IMn+:#+,!#Q̰ASIyL583DaK6q(XqdF`ʄq24gM cLmGJ$_$fjSN&pJ qJeR&8xLL\C3 ح cB+KuV2[g1ǩ0$sB ANRl=Xg{FF~]_]1\ wCvƓ};~ٖ? Gp6_B+/M|{~܄ծ<.6_] ?.>`f ^c/q Fs;>|l* JqqV_{&vG嫭,5H_]tJ+`T_ݢ'Sb cf >.6$4m1(г3#ID drqʗ,ɇrCT/G48>#%LKHS\؁#ZB}Mi%EDgq0h0(Xdw#h1nJ)cB ^,\ rbW;Q<キ+Ψp\$'XYUgF$6?"Pny@%"zxלnٷkDe "A1EiLZeMc7R(z1͡B6˽ain ZнLP`GGƆ떰 1[UkSB4S9b| gHFc>UXo&,`:%ya\f(sЁ- Ғoe{kɚ0Wk<䮃tg:ðw̡03ڪ;ɶ`p5Y:fcmm^NgI!S R=UNUJ $-ps qc}w\Μ2 .1fv_'<b!r/BHՖggw(&-Ϯ9̣M{C?1~AN\F'^ODfFPE\H>x{7V":ɠ/IwM7;%m2 [x30H1r 6EfVBnξW0CNA2AQq|^ G*>A-xkUiڮ .+oCUsٙl\`m>?lfU*^9 ] AR[Ƶ~/܁УU+w"Au)|Չ>H[d ʼnP8tǴiWO(pG{bY!ka(=vh>NiLOZ i錦2y ~n- n*x0+2 Cc/f̴)s#;>[J0J#xwhl~x 6db뜊-RL95>e(M bU |e:Oo)P+ͳ% ucun*tFcUmr>%a:GL?_e%4:L_/¸ a  %]ԛ4|n$i~~w2cM&x=Bn1ƞA7 rtҼ6q5lväi89R7V x$%I|D4&`-S҄"϶'C'P+C`lP׵y$Ӫ?hvJKH!mr)̀wQ: DMc'6$K1G$4-=_ (&T_~7F ,!@3kNi؇t8`fVJa-^+h]H>_VK%4Q(űf/_2[:;v N9؛Pv.Ks))ؾD`1aJ-ȧ,HЋ|w*`j"2 hs3R<o\Xz-\A* [ kT*=*ՑoדI YXBW\`pTD8 JY-\+wl}' -TS@-L2+MUSզW =AˋdR!\#Q>@d7+X:浖 u!n%4,XAzsl9A]Dg^X賹*|eLƉq LPg5 |Z }3"Z BJ9 ]:d, S`VW[ΉPB# 3m?,Z՟v6_AWB hKht2&O%]J*"F1Pl lIRŭy9$rӻl51d@DHP3^;r~6!IeD W?3M$k+Ql@nQkH VbRC(Ɉ4N,HwHp1$W %zb*{_ 1NL-W"Wrh1B5f2WN*|7 2U!ASTzcNyJp?0Vf_H%D7-|힃UTl-|#*aZDo}*\Kr Sx7,~C\W/ OzAfa\3㏖/7.&~sV`a"b0LwѯhxYOw;{80F"P%11ʐ[!+!0 *tnEH`|;]t:gsk0[G0ü,?堓 kv}bӢmᬺx >Nm>,]pth{u8su0(vOYz;7r>^tVIk: qӖO?7tíX3<Iq}LPT󅊿3mrcgnA]LמR?Ԕ15 ˦Ato^]= Q*'MߖSvu9E@~?7^=UŒ_*nڻSZ84 -/h:ךZuQ!f .Z+Y9'vjqP!duï/YZ%4:& (z-W7%FlqSA8Jh'Bs,Ynná r[OJLI$}vUZϬ ;jݪ^ZPPG՝ջKPm}q5JZeQ%%q:0J75;"e'g'paP{N>eѽ@!TRt[!yKhhei`k;"h{j!9Şg:JH=淓l^y'Rn7 c8I<C0ւXJPCw`g&`$ĪXTFɩyubF+},q(l\ЬWǎ VrT Вj1!'(U^sRchMwVn?f_Kx < Za@lqLXQ۱"M]{z=WL/Ûo'm)iDyEk~FnZ!޿چUxwgW` m25ê|YUAC}_$s~5cy>& Q&Cv1uDp)lV5'//mq~p<FOYSh[4!7ϼn пKö$&v j m/)ֲrVP"ZJhtB/Sv%,F5lm/AM2-iX3&(eFSx 7 !!ggBm+yёz 9FX.fvRGBւ|/Cr5* On>.G`q4Gi2^,`Qլwѯn5e)~J4SR} `9W:fɹ8OUgp4a% |[ d`|}ցzT~B[>o[P~:C;$l:.>QF cВ}DqfL$iT:7{gX7S]}^1 !YӺ }ZIVyܲ.g5zv"Iz*(Lp"FHq 61)2K_E*lMf\/9(Bmb{>CXէэXEݨa)эϔwxq|?zcCv}31%׻妛qx˱g6…R?gsUu+ё`&m6`4Vo- %(Vv?N50 Hnb6@g}PYGjD%4Jx p%FR E/whR,bBDG1>ք ;*ȏ f 94w# ͭM Ih{ei ?s%4ΐ&,録fo\|0&`:Pd^ya~skBDoLSyKM~Ա,¢Jhte<-LUypÏiC@4|#)$+/wiR)"43w013p;1)TSUC9]TVIJI?I){>m;Uì?pZ aű.|z@7֡K@?e,Ձ-ztwnAI@5_Eu_w͆XѷL'C֠z̉L-RemX<4f̹Rt^Yۄzpi:jD0[Ռh5N,w0P?7Qhlb۰ P{+aF=wo"؀4Z'?4Tn狩!: N)m؀t$>J֘2[@;=F)9o]9ldgi2J{>餤FO'$ 'swd2r:ՙZ{O)1RX dߩl#Pԙ^[ֺ>IvS5Eؘɫ}.Jh5VOuvh!:$ޟ]=e@f*,7 >hmZ3{j3S;GgMLJڨ'M*+ZwfS}ӉU"+kq;ùva0=sb^sZB53hɌ:kv9SBvņ56vF<_eK[euX6>Cy;䪑]k#÷egߘU:GyG?6ս&<$YC|H!>a>gg9_G&vلySf?.cex&mg\Iwvft|t|wOEd'g9%.uUՊq;r6('%^e-kF=uUlOkfrcJG8>-ѲA:FVHUBZJ2oXfV$;Zk3KRԖkOplvxWHg'Gzpbi&3jjJUָ `5㐚vX"!/󏺴<:,!xu.k5 c[FߨzP #$!`g:-׸LG['\t\tsX$kzoT=KʶF؎#Y?z 3~{*h uzQ (ܭ_|{Uyz5ъuwfly ,0.RE4DÈ!ҩָVZ[(WE'_KUkye]= &Ez8zXfYw.?a3_e9n+0 cݝvbwq[o~1{{qrЄBWbQ;HN}2Q5&X(rM=,Y 2;ދ-{WHA+i໺:1ޫ`tjC"icHG |~=FAaƨ9i7<31cֽv25,/Njۗӣ?~%3;IG˧%>d1?h㚻~,?Rp["* %V5 S"Qvoe*wWw_>/͌;] ӃvV5xч-`#mj%#6@DehJƣ*o:\&[JBä́&$/wm|amsmއ]0{#.Nʗ-lJ&'EI SdiNJ\ Vٰp0A@*j|q܎Aə" x}pқ_OӇlN| h"[ZU1D-@1=y4W}>1r[8,Y8zRkmH΂cQuiŶaҋͱyuLv;/{Rc;O(ב8,u2;S]ݸbh~R=@OeSYޱ F~;%?l't۰i?6gPE*K{ k|F2;̘ ua恵ny,)cCt^ƚlXŏ6ApʍͰi@6 a4>|[h5h~cP} ;$LSTG36$gq`px=)+uI:4|۴giZOSszJDFփcf+hiH>~J_jmFV:?vDS|g&e7:;)q\)%IkXT|B՛Rd_Ci+-,utoה0,).y*Ü>ݫ$HZ0N# @|G˯W1#mjB> OSL<%J!*Rg Qǔ_Ɯ7%vxjӭ,e o5 D/(#+0hDF#0Qш0.ʡ4k%fEek!仂3芕$#?fJK%0ZHҸ^4рngX :բt1 -F6EDPʼne]hCd4x4jr*^Gtƅi*fȅtY I̠18l b2E%$'bq6(RHPf!+z#-helN8Y{Lf*?Sz*VYSfKd#,FKhi&˱ @CO6Y Gn9w PX'LFE"Jo)EBPN>6ȆE93ɌO;2&YP9`,MmE.Y$HH He3Ov$ 136l M"L QT2$$des$liXR!Gه,,#Pj@ v} u~iϕ?rF`SbMR0B%$-AwШ]X:(ep]30K1F%\5DpP$^4n|A*ڏ^E{;\;_IwFܭEg{o*UyVo+j򟻌2<\WIԨyS.fVH}=E=uN2s1Cݠ Nx7O-4[V ;EI#L rF/m; ʐ k ilRNe F3)XQ^ Hd=Ef(ˆ[A^Q aMC%5q#O'qy_VMAgeEM< (Qx*N8欛ؽΠjAS|&s4xkX^V(VːT="0MayuN+hMpDc޺&j{&@DO7X{gC5($/U1(?|]*v g3SPQ>7J ,ѨEd bĂe?^^^'vb wi\":_RP:Eơ%ڒ a6[zVu׫eY:ɈDz@+'2c( { ]+1T((SC YհĆaMYdž޲O7OubϩN{rM5Չ9~ckLoXEy˻MubSfէumz z7*re}#CZ]E s[և =MpIrc C ćVxdh8=#s޷Stg᤹f^R.:LrҊc5;?5>aGP0jdS4+ObTW## d^)g "RrX;+7KU tNY4\1K9ufۛ@+Y:[%lv6& —%hX s;AraP뤮v)l[%c`|AD 6 -kcc?‚T/](%yd˔RDk7>(mH!eƞحGXȧOOC g~uReHHIiw.a3}X\vm,_q>Nl'&gAXD($a͑Sg'ph\5uXܸ^4a]>|$@.ߠ傣BK*hh1q+ai6{%.9E?hx5O.x`ͫQW<ųz<8;E^یKk2յxzϖf%U׋-.wyg_/ykFkD?߿]<5jog(Χ qYwpo bY0y{Un+9%,p,=[U[f1@J$ӔT[l-ƵS^|5/N(s\,N_qW͘fwfECnVa4Og3ze1g63<_6wEYJJP|ߏVmy=v~s j|;wOBB."R7Z`uGt7zrLA&VO8#IDz;dyM[l'V3;Lu?XzWZ)GzXqfEO i}OA&>ߝ {;# JdL%vB(EpF k&DLʼnj,mF] A# ݉j4!괘=wO/D]] ޾m~&|g-uzWr~}\T[a<_e'/|Oթlm]F%cU9<}p_F,b[r0֒3@-sTT𜴊C&9n%`ʼnmzv 8lƓA#;Y9Ǹ&IsȪB a oixdR\fz@x\(`x#ec&h64a}:XhPO̜GO0t%A-'.wm|˳BgrO&ymC8C4 }aA{=b 9z ss.9:f~5}Еk>p6hnPdJaeOmj 8(2b @o/[ԙ.LO`y㝇m^gcP+EF_Z@k3A^ i[<HkB7'rB9hdw>L ؅g&=y ~-hgug 74>delnDzVYVQ:.ٍUdQlUgW*ҋɓZc #jNeK"^F 6c,z'~c=Ȱ2b2ōe1L`F61cv?X-;b3;dfb U#z^)yĖJh$KFGFVJ/P+e}df.{_|Q\RP.cu0Ԛ[]ʻGK.-w.oZFNĻtumhw-iyɅZ>Guhm>ɼ*(UwGh5l|v|VmX1i]㱪l7)cMJ`WժVJis6fn:BNkO(b&nuzۜtb1*llmVJL1z::`+F09Kތ-ZDCϾ Yh[I0bVgVsu~aK썼8|sB1H7@(xm͸мp,Yo&67F=otT΀,ZZc7ͤD*V 9ΰfLNg@:}I_wv0B)a:$0}Πa2߆ʝU|z9VeA76Ye=ţI97oֈ*d'(%f1l0.(OGoska@8y'W DŽ+ֺ#)B'uvᘁ"xzrJԝwLA;rkC!ȑ[ yلBq˭[\ZZD4݊>^/Fd(%eW(YUgh)j"FncI:D([M XTW'ٰf,I7c?DMw3UZB?"'Wypl~Yz1[uqNE)F%7K }oĀOw!}FĹD\w\?|6fknhZ Z*P* YVTqŠ-4Gf IP,d BgY{޾m*8 ,D:*IAGHGf9YML%ڪEL'6дKk}uM$yN55Do+W!i/4#򚓢\;4j ֩Y1;4ڗX"-mjQEQ2AX+G&Hq]+l-- `8N\%aK16YXp 7,FQ`#Dy[6:QNN2H=.눣POpZm˷P"*bnN%( ?&uuqhu .E4:"X[= ~?/βȍ{1twz(XJn?eP-j q1XdA0 Δ> DmRLp+,d4 wdc+%-L:p-ҋBk5n:l @ؾah"N0\JP^ 3S؏֐q/' ߘ~F[0f9 J LG=FcP~T/ۈb`xzld ÚϤh2 7 :pҸmxύWƀL&y6NL=oZr@0T s喬@ WFN AMd B\!h±@WF-ֲpePA˃SJϘTΎ$Mfa}ϷK c粡5gm#"&wrףaf"Y!}Eј={$WQUO#QbzPbg^nvQwC&3qSGlIBE9z (&=*>Hi1:r3$aU,\a+ [Rk&W\هZ1 z=f9, nTemVVI*T"`vo(%]gP ɔeSq cҎ]*NiP)/#:b^r{/R|p:>dk),ۘJ-1Az夾y=!x}Z-p؄Sp%[KBikxx*}:oB!t@8=t@2b-0yN#wY>`v9>Ixx=xrX>/F ^Xhc3rP}f"cJű Qexb:s֥Di=[̴SB% GPߩ^{uA988PxԍWSww IՆ![zJ:sc S!nb.?dg8C ^3Ow"ߗr"Ozg^TD ʈ x C6`nb]oF:P]dROʡ>mmi^_Pųɢ{z3)JRi N-ǹjVICLqęeg% &Ĺhgi `aaO 'S(>]ځ!ƵSX(yrڸ T5kVI!:> :.X.~a{ ̔# vOMuQ*󜔔Lξ;ZPg{ac{7?jylwpxOK6ﺲelk.ټ%7ߢt%SqT16&S3d E+~Hx8q #ouݤ=͉34CY2㋧NSmB]+ v@WIoXkGM\^qgB#vCC3G< oB]TK<".Enwyݱ> ef yۻM$G0W ;W$aoPJۛCgǒ"E`k%O`_qx":X\8ݪ۳.x>"ݽ;][ggXrW8SEGdRՍwZ۱ w<:Q%@ׯ3p٩3Ѻ9ËWpqL7<*a`/ R:(S IY뭰ʢ@:DFs:\o29*z<6NYO-^V?)dĽE_\z7VYU^}v?޶7{r߿+.]tuz<,쓏^)~.aY{Y7|Χ*c:}S{}|QP^\<1vAXEQJmKO5cYf:,m /LPK٧_<_;}2C)h,g,qV7Y~h^v G u{_84!RC{GK 7 {úIJ`=/hG-u[YݏIsDXm3{[BSb=+>#iDՕ]BO5#e[|e>oKP$% >)KR)i.:"͎H́kma= ĭOӯ:"0e] ɥHl!~E|SW wK=>ZӇ6-k&XCa-o+Lad&׃4zVtF pM0Yɼ 7J_a'EC+öA쪃QxhRlz7b/6UZU8PR)aWl\ՓF0؄YW 絸C= qkwaVv5ƍ1q|hfΎ'܊7 dAVNucLsQLIpB׬Vj5Ti>NB{jƀ~"\01t_]A$v58 )&CpLNx- ^̲^d[?{0r3SsSrSh'jqr} q29tjڥ p_-W$d5*HLl#;)rɁ%h(ylXˁ*攳hdVt]$\jQ2 Y|O SKPaۨ#deW+zwѶLE-ԏIcq=b%M+dE9:t_Qx8hX`SW~2ar>&cnP 6/E;gvɀ'XbcyO7U5B$&nSc/{[?CZ9!o17wU`$޺g1LbX O'LkJ>^|6Đ*QJ5HH#fVM4Tm) =|T%$/ HKRb4Ajɥ8I>*112Aum_Z5⽯ڊ؞HV&5.z-8HAlמS՘i*DLTpxwՓȩ OVR*^$Z'PfݜmS|O7&vR7Ŏ4^b2b0t,dP/)usE܊Ո`c!JB>t =^{j0ion\vMRU"CU 8!zjLk?&p!5$Uw(Y'#11qxƠ"c ; '~~3Oyh_oADOh!^Ǝ8gË7 N5tcԛO tmr -5xbIh jp$mNXBj֢CJ8 b.Ypk80 'A8?lK{EbnӻW=0A\"2LL}jj@^>J(K z5( Ӝ96 "K:ł9E;t;W}qjeDidYE)7]ݨP&8?|7RhCxBVۍD俗Òȧ\[B2Hqy7^ SY| fz=< a-4ʔ- bزQ qT)F]jBGXe¶UZipT8f[#&iڅ e0yT.k詚cfOyY9~Kd/ =WjDt=ũ18V'N2%+Z5+`z+^ `LJ!؉-#ZV.&&+<sL[n.w??{RJv你6=ߠkǁys+!y,wu^d eJQd5c^lrg_Q>ລ@Rqo6jk5 a9l2,yY'}rݧ~kQVG/ݮ  {!6W s8,6:c[^AE"Yum}QNu~g\(]{#k:80ӣ{<|yyJ伂 ߶P;Zu|e5'57ɮkUeo}tn>{S#0hX?ʟ04(z>ނ[0eH % Ij S40u'RyE41TAZ7ЗF,i(nx O?A=8#r6\̳'gLoGcѮmKǓ9;lHDPn,&5<#z@u" oeu,v܉F)cWz QK%f+UM iSI/(qu0o.9۵~;3M~ص\߽J2#DFFL <ŀLb%Pd!@]RJ54*6vᒱ& Fo{VZez#lآ3RjY+oj磤l-<־+Hg1L>0S(A1osTJkJ@L)eldlMQ|Xyr i@:8)A138-`ãko_$T9|Lױ;eN~8>/G@-SM mr_Տdj:qSaaRWZCJ1?:q܌FYlQ!fq B(ywPC6ۼN> Sg[?{ǭy[ 8Þ-/X'XvN_ű-5=F4|0Xt,Xz@%|!{ۥP~j[ۈt}U,MaG< L͆Rz#orbDc' t2#cx3#zf],vGG?էp{ĞZWT n}=PmXfbС:a)վ7,+lWh\15nM}Yͤ5@Q/Jf2XZr&Z(VV⢸~IH.Q?x[ m:ro9%EK~Io1M^)s66 㼉 B&_(韝{o<`Bވqĉub<*Z[bkoi~'lF0֫,řoz˭`WOn#Z{Jw!NyW s ;6lyКnČﳛ#:}qlW8snXu't<?|iuΨ˴st}Ί37(T@|#HRM.vlT<'R*ʶCXq pĐͻ,P{!rQcoJa̤gP: 5q^+9ui9|?t2[Bq6/)'+Y=~$}h@**H6ƘZ n|0坵'#>nBfށ PaB}B>m9"}ZKřWi_F"Pܟoʟ a%JD\7^_ܾgePR02!Dp#սnM(m6=ՙԈ& sRىIcD6/VS\!E2'Lk.TUl Jcp>v3z\]+m/j jzC^9ҴqR~rpJjJ4?C1;=Q(L),v?ޖp{^36: tSL36g!˩l&ч/;X?gmմ6.UTog#(=>Csp`I4*(Tr"v )ƔG"uGB1C$ŤJ~*1(1 oPPJTi3t\ Ԛ2hle1FrdVeT bjkp~jʈ=*<JnH0*`]J1&2ZLp֖vC-+oM%9FMPjN _;v|8\xfwp~s>\q?\ \j|'h,8:8j@YMtT4[ʉl]rތxگt'su;WFtcX*SF%ޡ19W+&O6mrr=fY~ta}4 CFţV+zJdgcɈgSO]up'9n FuyO7;{"w0<\)}2V((5gkƭk_^\NX| !jMA38V52輮[43V=g}=>$g͙Kq>z/MGCGH̞5zlzǭ͖v'}l6FQALMv=\X] qÂ[ysҸ`zG`2l7*WWvz^#ǓЃ!ںo^JlYLU*Ax&<.jyuN+[o]yw?gWo?J lj6T,u bS fn~TS<7^8\ i"]'6ʤy 4+&gd˧ />ݧ X¤桀yH>}vu8J%xN Ykiwf!aOzZß!e]]Wz634р{cSrhO.@lN)[b8H8%jPo`Y#|֩]j`Z$?k[/Yюu;/6G(?ήt_n_r|sާ6I?mb۱3mߪwmg|? E⶛ N qȖW$E=˚gFr Hk!<<͹ %ɅXZx+`^]'GE*o-(XC`AYNv ۯcF+OIF4ǔk'%!8I1lJ$[,<WnWu{a W |5AkޱϐBM]] ƒ_%djY*t"_'Ln!x 4̗O޳koLS2M |u^wkVW<姁 oa>T9uD Ydfe@bqoG –Z\GC^/Oz(Z0]ݦW^ js·_(#&סy+?9"ؑǣ%H&1U3YJtduKv x&ln#]mnǾ߿y7#(ǵm9 _d%}y4trI(8j4QA~c`&˕ WF=)Eި%?x3v~1}Kv@OMZh?%߲ݨ2KwR zmAf-hSLq*Kӧz' xRyuwOvܛ,ulo-Ϗ_g$+:\Ƥu܊rb{Б6܈{Rҝ y[54 -Ѓ<gD,m$X| (NRso^z:؋?tP +zffǵ߻iY^ufNrM)E Õuai~(K^Jc %, . lNf쁋oN_>iް5o $R{R8ML>>q_U!њw| >?6璺Z[={OoUg͵If'wqp}oE-BmWa-u5Гr]G)߭D_1l8L)EgW; A s /3IfVb8J0߻%4QTl5Zg~?w ab/@+ SFNtaenp0S6J!435$8A51k-%qE\ nnY-fYH̬5aǂ \0$FZIڑmץ6D)j͹0VFTib/'CR(Ԝy7d8ȼb$rI\jB,Hl>5^P¤"3 *p֌+X bĨSq(%1'a3 t~/a]||?,"&3t F.fqJuEl),e&ΌujZSw |,e 6DzO-Ʃfa{2꥖+4Imve,S1CwBI3b2 5$3GFbePDbN݂@$K?;iG\"A.C`̓?Vi~ "U26aW_/,$5[Db S4K dPj^}= lǎ`fgi!a*xq*2bZRp"r8^h<ۛVf=P_Nf-JMZ@>)_gyUaf^^5x چ !)͏vt;tŕ?Og]4'ŝ8SA;fIxI wԀ/V¢L[8ޮ rT~śNj3_lF"˯ķhĞԁ5I8zlW8NSNe40],N.$A0g(C~)9}q7_1<@g?/ծ } T@{uK;pII8uUh "8Lp 0ʂm~=+?B:Ϝ BqF#BѼsQ^.qfv.WTKv~sAO}_"DND\("5\J$H4UڡJyXBIB#5مG]84]~SPk-K3΅$ŧqh&DɷXkPzј+ǟ=4&db 豱bnS8 .m8A=x|J#OyGT#D$#e?Ӂ֣P}M껵dTM2!Dj&;K'^oRjZ.ꗶVgo\" %Xq*V#"Eۢ?FJZo\Idњ!T!.`)El M:9LϔϚ4'B"["oͬJ|7V^V\MhsĚu} W(Y[ 5g:ő]/%{][_{6bZ@Zڴv h^`W]ZXIZ*~=HRl%R\-VEQi Lۦ趷Z* xsofź%ՌCJmӵQumTQ]k^:uJp`T8BhpRZAh:+ nv-@- g(&!ޏ{H󝖱4'\'/9/d31Ioӡ_~yyٷ2}F [NQ*r|Dg S}&uJEʲ4Ui*mF֠ qҝpOVnf%0VeYg D i: `ƴAɌd{epFT<({5'l7Ry{ܪΜ .\P3>ab+:iXP)/#,!#V.FUgT3RmSй}b Vwb暋VSͥ"Jd.E@ g=/@0"8ύ\@[a""(1 K9 +dj{f2\\]katf߃U_tӌLqYZ k(8Q+?\.ϋR8Ц'.b]q!]~YjknGwdIPk\]94DɎNCN&ԓrfVIwma/~ay;+X{_b&.0QjW7F]NH؋”u8 ʮņU-CL~m6?c.NѐWkN ZcDSv*,#PwQH4b(I8W*a"u1K:*rMU ǮTʋVjF/E4NF 2!@*()LBw"D'kqGɪ\ckWsU"߭ 5WQz:t)XMwmmKzY,~`.Ea}SLRSMRҐ IIc - W laoF~JjE=~_ ㎫0 m 3&x>ڀxPWot1}mc`\b҅Vpb#VU_| xxqnMGq^tV2uq?rF{O:0fmj*|y2b(SGmgO[,Ē/3p̋=(Ѫy89ۢNk+ъ"j̟Pt_ŪhX{vbL,S i8Til)R[0EAZ=cmk, E_&n1?0t6pӳ#mq|-@ıZcucm MW: 5- n6.vݔa^rrnN)bxЄ^li 6U(`9V&XaΘ;b ݶ~+j5{(B%tw3#%О6G^q;8b7:[bjaآ[ xh4`Nu[MTXw0{ :Rptxte[9E8zh FVyST˽ÝΩݬS-4v2$Ynk[;dāvJM )>bW dIi]B`E rQ皬ԱۏoGñBj"KR|C$~|IvfS3ixNB JⰉr^|{8Ғr3MUW|עda(]NJIIwr֘%rcT`&R&]` U\cҶ-ѿK C{(Q9uYZ9E*vLj>qsO/A|] bV#^{X"D6ieOf|srLΆ{,Uk:yr:"_kB!0i$Ne9ohbjxye74x*6\`YQquuN09Re藵KТ<S~(!5͕kYf"]:.P<)]3]~xYrQ{ ISZ'0^>}~5$lOC+قO##܉VswE3  cRZiJrQ_ߙ =!KG`6g_ @g.d`8kAJP{iݠSK՜W?YmLLZ!ff?nFY/A XtbTWY{,6m0 г୵l]i>]G'PpƘLjQ\~֘>XJb\JbbLd?>|b+r1qtxC/d\~5u2S8+.^xi8P'i&F`/Nor6u0*{tĄΙE[I|VXN;( sE I@ALR4Ur_e( :|wpESx?gsyut /p*@SFqs^ їV1lN`QvwO/@;lD$i%P13$>X%!Eb &YBi69P=hWs:ҡ<KP~JP~9A,jdF=ӂ90< L)7^#TL4 (!KG݁jIGkH@CK]/~OPJ ,^qNFjĂP@m 1TgHz ˕Ւ~ MJ!*+ӕк}/bi0~7>ߟOγ꒜x~{G K~󳙙߻V*æ{(=ae0 YD!DA E +` JRV цj 3@FNJGD1aF/۱oP-pܜthD f_4OɷY,=e1Mbd;t*1s:lK'N+# 1a@iR e4N${aJ1ZpMzj~d9Q8 К$CxC(<h=Rmg@HS2B55MǏ:#k+9XݠSjx}pv5&25)7x.Uog3EbuKJ-,mE0K ҷT%H0:&7{!~<8~s8~s6xhK[ARNS8)%1gBG$2`)Fq"5˿;P-P03tL[dR-d*TX.9W"J>Cj&)j96HU806+,clˁL_5C7SBƝW֙"#`B)ƶ=:ȨU` `̍n6,p8D ?-&W_>=U(",RSh]GXVb,)61" h3IAĭ搅bqy@?Uhȅ 7 l+`#ku 7^Za]` ΥcFq̎:&9=ϐ!?,lxnwh:bypP"%)('ECRSOT`DRm0P[$ň%,$( [ZGi[&h[7m5"Dhc*]sz|ЕdkJ e HHi ~5 QjJ VTMGr_m5wDeK "k:,b3|!፦)P^FhiOUfL()%'D!͍uTGJ%5תWəJ<`M:JB$zD4AKop()^  vm$q+ܚGQ҄zkQlێqtp&Jfa^kĭj$ХmE| en[Vmb֒"-s$᣽X721n<`S)]~UiV%5(aw0Vhf AH<72p.) zR{oi-|-9QĚڻ5ĺf#J#RREtD<:O[0kW;Eea*C  []]A0:i9t9=r_I '+]|fXt]>7 kcFNy D =՝֦{Lrz]Y@}ű%[+j<@%r̃%32 $弍]>Y '}웭t[f~V"eRQresC14p=eA?i wiF~zi>.bXS,)&sF %xlrB|6a*Z2 jR;g܍]^Gp(G ^> t0S!{ʭ?Y~"`U5f"*`JGpjnhĞt缸\!gŖ|0O.,my3 ~]: :6GkTX{i|Ht_V҈Uj#%(m;:- FMUոU򭞬ReY߶ݗ@">9;`sԔj:՛>#LL8?s& ਜ਼Z0tś\t;3˱`]U4lU4\*URjQW T۬X`kD - S`dY>tkYT3͛jv|W:\|@͙τ;AKTDc:F""1RuX@:ꯏ *d\9:64V3޵nt9)n\:hm3C8CޮRYaYRQ[rE# S&bR)!^YEtq)upJdgrT9G԰wjT"c{5=Vk&f)˗UCic*PEk?TH8m$1sbr2 k53s ː`~Ѐ:9YrĄa?]au̢.NRIj;u]hM]?"5^aL L,؜!H`(DLr,&%u Ӛ9bZw[%HA_kInDr" dHv26 /!@TػeWyM"O`x%pfQդJC41d,QMںpJpDafJ0רBn`ԇ"!:V}vPm(kxo*NB!L5ڗ8pv'P2z\^KB:JD2G%ԙVn<42%u9%K`8a,w0D;>IOWp7$*/ `@*B({(j1,l+$Ó*ʇv#&F(#kh(.2^;\6풤wffu>ݺIZ+4"n]##zk=nWX+8h($m1KX-͸J#m‹2HCT[!`$-TR/4^lmWSH41׹7ǣ9 6wf֩DbWvB06 ss\nGcxNv)q[kV5y9o}Yrqvg1rвt&)Q)7ǚ9G850VF/- 4c$ cs2#`$5=)g8-#ѝr"^dEwɋ ΐuϬ➱z垕jݳJؙEو28n-% "zFD|DsU KFDy5 Kp]ucΨN4BGWyplVޅ]}S^KXZ}j ߴ˱V=Hdh:m^X^$^{Lqco s*faWEҊzVԒ@6Nv0'(6Y'6~r²5찫2ei/kMR|@2Yg;e6M~zf`ts,ernN:)D'Tu;Z"OF0aBJco%F0A\ZL-o3d(ݥҜƷS0pcx$t3K3cq"Hx9䳺ftER_f[{=.ꊾbR 3&Ib80?L!HM[D7lI <_;$a4qI:.hDT⩻4~Ě[EQ"|JT.{N)nωh\ͮKXprdM:xild4fp6ԧ~0'/c0_;VܱDAl|w ፝3N@t r>N c>ފ^GLITTE(DPzD"7STf`3t N{:9=q|3qN/ao"DnpD#["8׹ls؎ H;@'_=0EAױEgn.yg%!yUCKC{F 0|xtvv!ȓ^}Y_O ߤLѴ|_]/~x_@Yl2C>`18}1/w_߄ig_~z~ *j0׿bo_6t#4JM V8M2;j"&&q"\9~nzdd9&|[>2#Se\KF,Rs[Ηo~ERni'a)1'OFx07;0@1>YTPœpaDRdb;:"M$t E 3-e דM,? gXh/%hr6Wdo.NxJH+ 'rOYLꨛl`KWg>Oi+wLf߯φ]޺.wy*Y[sȕ1YDTXqP8ՕrLSAf`:5xųvݍR։dH*QD0WfE& 7}D[맵 MI]LNݔeyeɼu_cJ1+GIqK@ILn)8()TʹK;ާMu}")m ")M4-,Dl`[$ GZdAB8]4DL dB$K/`yzwӋMvLV}E&`d>u)>T^2OpȞR VK$%9U;.1KG8c?j%vpqKǎ^_/j©rOGs>2v}qmBsNƛ`m0Kאu&1g?00d"W"k1lrS߄e14 Õoklԁ 6h8|lLC<R5I=@Lkdv894X qsI,2 99&9䥆nŌB}~^ ^ /qYt>xfgS^ݜu KןkZ+Roĺ˃1'Y)]';|:ݧW $}«T.."DefA^#sQ+O6~S{0BK(v Zwsa(yz\: KqXhM[~(U9fFSr^v^$ *A "):4q; ,'s[&8Sxn\hV{XqXGΞ iPL Pq{ǰΑjϬ]mW{ z|־WKBarr$H ̃=2JL|] 1ږmKڈ.6'f៸8I)T&'im֖%̾jڒ-ڒ\ξI-Q  !gEepj{d%LBkKZ?IgmKK (O,i㒭-ڒV%[[%K%GqBw>vC1s:,1{n=2[R#XZ[2.0Of/lFZƜh`b_;cX?%-1/X;)A BRPƎBH[w]^ [w[wWՅ@>矾\{"3 `2`Dc] .ܷxgmIڰK Ҽ$ 1qS*(>4r0ZU1?@>JƀvXnҮ}Rh:1z6<,esOf][䖃TӃ Lph6.8XBn vBk~ S- ȝl;ˇ.A)]zŒGnz1ڟpXTv˜q~Q'%f/];XqB O:sw!FsY~833fVH!Ϟ4qw Us~r#@k5m{hn([q+vvm^>FszL13EAw;{>iG^_Y·Op!Z"jnx{$f^,i'A;u$bF(ӪIzP+s瞨U,,vwZPΪq;pV4M?Yl+g Y5>&/UR*3+}]Za iőͧf!n5<ՃdZR|l"x@ax0+q+r"jtRmnؗ:zymn[*Ow4~yrxdq#-ǕkNhm<<݄wy6Rju8sPwn/{jor6PH+w.Pi=R5NEtL u0^,EER`1q@Y=3Zfq2SRF(Z=ޚinEE֮\I!$y {˜>=W/8mQmF‰c η#GF|_W}|^t *T;UaJq 鐨Sqya`V2CD0(g08Gs="׳W/ Z/GY]I'}ɻiW|`ړrcrN B# LO" wem$Iz!@=ih{Be[4$n{0}#X,VV HUeer5 pvHn/u~`^AnVvr+΀=~쟭F_  8ބ?lQ:Ũgڤ*%[(d$#.f ZźW DZ sz}̣@:mc}qMH'բ!_ x ˷yғՇQ,)`j%z`֣WDڽ 5huxցCe.mnEmsH;i ȢbD9ۀJ-M2 xD1O#4əSYV,'IZٕHZIsL fߚs0t971J*&dx̻vڌW, u6Cd?/>o4d0}(jEF \BoML TlُB2wie \bM!J 5`PԪ+c09^Q1r|BpX LAm;3O[Ii܀_6-ԺǨYTR\'z)Cl6sP=cb gh0[Kgg"AuOX+v+cF.H.~CN?t*!E_ĝ揚bwF+WN:%A۩j)UnocT5Rasw1۩2$^ԝJD;P `½4MI/@RjQRʕΙ1P--mRY߸}wo&RmM+[,ҥ?"Um~,+iKyƥ\RBI/0S`ΑCx5d>Hx!sXioZN>9$#Y;ښKw%Xm\%;\Ҭ#N)њti$B0:h a` KTg ;?}W4䀅a<|'7HdwN3O @85d-2ZNSq\RV QF% X!ѵ/ 0BԶ&JS[Mohj{7K}m\hV2ICPVH F2XbNQl%v5ӫ"TqzqIӸϤD@TxiY2AV CvzrB*dNKzIM?/i,\<|0LImڏx+Fԏd:gJb e÷#&w)d3>(-2ݏ~zw?8Oa8M_hH ֣!1/rv>pW`j_F0>f t?NU/nCan65~Ѕg #di􃋳v';S1y6z:C$LDGI"!Q(fƚ %X+qrƒf\܋W `6Pk(r"K-K 6e$BSiš&VKo=.Ayk/Q/*-h,U-^բDoM1w32_~e?`ukF?#\aDkouOJj&XQC$xK2Z/96i#{A4z)*zcZ~tR]i3ǔ}b5J:V\mܐ>$xGfsnPsO4/nJjp5?Q&+[KF8։Ma$3 Iʲt*$x1Kk2RRKX!JHw*e+O`J :n;1uE1Id1[wzI겹0lj=ؙrcѻZ٢\VP*B4j<@ kZӊJMP>2VH.sҰ_Ik6K8eqJ q5 ~#S|6RVY2 +MY*} f*T謍uR ,f^ECy;dratpske=hvl *Y/]] ={qR Q+MY' kj0jDT~W0(> YFp;0+fY/Nx -vzi(}eiE&WD\b @)Juxnd[.pWOfKOifڨ~!;8y /[-kMy} +-)+.hIa y@I-iwD]ْJ-)BAH'֔h﬑b"ɼW9 J[jֶO!e_T8j$60d9ʭ m҂")ꬂw9ȧI-AG5nq\8|o}!}߇1dTr#PE*%T}^S31Rz0IL# b׆yIaKotd@ A9W^cbd`%MD8^x ""+E@ iЖ$ D35T<DZj]M?8+["Iw1ZP0)iӊ5T;.ź0Ғ"`%MFiiԠҔo6',KcS5H)W#"B #  !!4V\{䌧Lp` y 8<ɌI)㮫 'JUy\#BIGӫS/hkV,)Iw\6Ẅ&Qt(嶼ٚ7x彑Z1縠ZH΅*/RRЙW$ht)|.Vju26ޘuT+MHJ\)Yae3ΤPp5+c}tfQ0pR"AB3=SN9-(:g:I/ĆEV^\AV/PА൤0 iF ]։SŦr 1)5QEN!?LʝϨ"1抮\D 1 *xTk(0$4g|Y][LRG#_FE6fL +bȤ ). Yv|  rn1=J,#ⲁaN*Qe~Ykbx aF4R0~PP(PedpIrzܞ5R L@G*C:<6 P63M'fWkV.~xk"|Ў/Z_;m%EC |ս0n*SOJecI[6}{D"8bȅD8bbC߬[;ɯ 0eW7RFdX] ڽؕF=6e QHbWNynZE 0Ძ>۽x%HhDl-[D/]}0R5G}q b>ӛaRfb;TӾNqٛpZ9]Q9ɲ7Dcf.D4DcTMu^q@rڙv!0_If]11Ѳk}SݭLo99xҗq/^Ļśv7P>{΋õ]YV掐 X0t՚5f$͔@ Ĩݣ.j3Bh{KvSUlt5jk_nFp ִv#M)bq[zS EXz0Ƨ{7#O \zO#ú$۽Ry&|T* lW0"Bjno_U)tqSR=+QV) =ÒlxsQ]:d/nFw$@C~tʡDfWN;:UXm~ S%dpLt*"hCzW:U V06;'pR3_&@QTWvΙw_NEd71D T1v#Ey4&Ui)&F5tOaN%k@^ 'pijQٻ6n$WXr /7!nm咊v]*M,jIʉ~!%ɡ4yXMO7 pcpkC+xf?`iXzħc1Y- @\^mzJ /DjsHcMumr ":Q9rDј+)NQ3J)tI/  #_m0D!^;<@?}?p1F&Jq(EFJ%ìpvqJ+:]+>=EGwyXϞRɫOjkR"nqݯ^]9M}Z-./)2d*ԕk~qiu>X^-(0&N~banF[i V'׍{}6bfstӫ6a ح\E?$< F*.~ڲύcOqy"܈b2bzTLBN70 6fr%l0ɑ9c .( 7e!?,dt>YVM]77Dw'sI5GBˏvqy3s/R\rv+qWk9zO ^+5:9hM!kYOGwX7Iˑ[%=Jy3SC}tp.90wZ\IqLYhQty3Cs2s 42&E/әu:SNg3b6τc^PB;I], 犦T{PN:')L2%7kB"@D9N\T4kPm0dTI2*z5Z )GDq J9N\ċV0FyM؅ۺm Lӡfo+ &Z zR63Xƕ!;q91A5~ :b5.>JÜ %APm6}7󿺾TD !J?b29=X5S0~U5Z3JPkXM#DSj~ SGnB ".3BIjnA\X)g;~W)J :CF)kRFCӠ4zlTAT)*_SZNѯ'0sw0eHFzkUAmXSzkl Q-fi[SSa ԻZS3-lzYq'A!2DR|RƖͻd1N_|1:&'qq䯳<>,ݨ4hT;\:- N7ErdCBu>Wɉ2ۛ7WPKŽ峩|g$hRC呸 & :F5 +i2~e%5!S=q`%v:A-,@KAIx?$ոjI: f?/OdG0q gN{aCѥؤ@BAyRx e勷/޹Xx ] I(?c͗e@2ڠ+3C+qqu}w` uNS"w (rО uȨP/{M vIhfcNڻ׬Zϑ O͂~̞Ny!5͜th1ZPbsrʾogS$+,_]^jXn3ZXňkRsI"\Yȍ(z?|NF]|O *eZAl/ L00XKZy0{"&) Nr R_wZg Ka*Pqw\DSr@wF+ Sx E|Azs0i71iEN8M^1` Dm&J wAd`R?ǔوˇUZ 1g:L@Bg`&F\:F=$?xZgR˲Y$g,Yр6{W1hAq9 NΥ$p5K.rιZy>{ԗmSzg6@0_n>iQ D7LJOUMر(V lx(sZ6ʤ$1I*) *diJJɃvJ$. 35KR[l+)qi 󥓚w|WT{l tG@T ')\s#)k{) ݓ-KMtOB]i&*hntԽ6AUz)3j%YLѬ3!SN=j4͔P(ZI˧`Z3.RiI! PSkNFՕLqCJ)Krnbh-[KG& ȌBY1( m3]Tg$> X͑)8K TF픵 JVㇻէ,uسs|tX_MZAkkyߺւH#ܠS77{Ԏu(MvHL":u_gk6ּ[Z_M'ot## =~tih޷,G%WXݍ$J-#{,?7;pJtP}Yf([2+#S%~5-^?T"e|6Qպfi/)T1ٰ]L.!^qP ]6h?OU* oBBS^h2F@'39 cɘW0k6U+WC7aMF^".g`[KQ8-A4 3ies?,CU.ꎺY\ca[+n%I E-n[wD-x½^ C/DjMBPfI h.\ob?ܯ" ݮu֎Z{'2o$ʵ{ёUN>z[[ N6X3z(SNnZ׺u!pmS%бzȇ֍z}F:Ӷ&휻p[ֵn]h\EO)Gv;W~~>NZU(z)ŕʀy">]A"rZ`JKiE")8n BRj?L2SR1 69(%("+IADw"e?QxMSK wET:O8R8d˩FQ/k_L_̈yzm-N),ϷJt q2D 8#քeG&,!3׽XLg%h)DW]8Rӏ0pJo%a!AogO%6V$Aa<øΣ(&(U/݅ ǨbLÅJN.ȂE"4tو5)M^ >>C2UÑt?/5Dj4D֚ŴeM͎)8cVzOJy1VOޚ ox\Koe>?{ƍ0/ ~1c,ɉ^g_4٤=Ɍ54N![uidA3RW*VU(9,l~D pfp [ݢX {(Ѿʫ,@ )T" F]gz9~WZ̜`2N'T`D"#I$݃.- QE\A8HCpj+%@┉!eЅ(Qƙݶ Χ\@xn|c!Bj}gqv'gކ# 5Ah Akmsc0e]$zhjJ y+?$chi~,C "8PN\{5!Wf# O&Sͬ[L)R. ϵ)l2+)ҌBZ 9S~ p 05BܪVLo%EY0V҂ (8&KJdL܋/cpVli#VpظْlrXJWTeq4V&M.GAҍ7QssoWC|rrbq㽓`,=$$I df$bx-1_^:QH2tga \KBChIRHj#p.ojZz҅: )`z+`I%@(P=Yd#~6\GSr (ӧ SHQ:nǦ*}bJk=F0XFF,}ЭY]]pL\ZхWF ^>WfWfO:?%IYf^5`ArAp]}ætwpv5ھQp&XF;e"8Lo:VY~B _5|֬1fDFYϱL$˾5ГS=1xۍ]@,FQ@if+8֑C‱h<=XHSB>!=08f1 pjXXϐqDl̞$MZ ˞P,jc}Jx20tw}Bkw~c,jdP17"ۆ8BQ7jcQz9iE %>b/-HsĹ̰2"`+6^*6JPR%3c,4B1ERFiגSe +62dnVB;v?f-J;7yΩm ~6.^JkM9[Fta M E 6nKVB!$%-KQR*)Bąb̅s.0X?&Or_\ݺ5ǩz;Lǯčw$n#q#=MAAJ .yTa^"2r;JBJ:t;slʟD= _ qVNsokYs`ߝm}[F|guuM.3zls[9r`/ 鎛r5:LծflzV F ~\Y[VyZ] ҉uܖfHr@oV{*@i@饳̎N`r1ȣkyZȣVbǚ杓$sNQls>Oޅ*7niVxyDf-A-w|Ѻ/hxTUd/qR>j?ۋhnV0, y>H}|~pj9ڗ?T̯nW>T~w\ $:M2L׽L btS{s(t{s{wQ7q GwxY"Lah(k4PC%$Gǐdٳ+k[` Kc.ʙ I5/J 5.e84LttŊQwK ݒۻ-/±Heu^{]& sETA~NG_jʙ|4%%I#,>Š3ۑ'˵rj PSܪW  -`L)qP$ \43n%noĺ].6Q"$k}1vt*:$?L?#֊|,+LyYP![F*Fv(f(,^ӰJJp4@Z:֑8 I ea^oKF -ҭ3IDm[FK$iwTXOOU3~ջWKc)s*x^EE%Not.9qΫ<}9&oy}᜗bV׏g7?U?sW,߭ǏyOJW4C_B0gnX9}'B($]; {̓oT"ȥT Cͯh(]bp{1͈*8*$ N0`ň,"V ؊j[8@J`֔siS:3mn77trYBDPEN!$9TI{1'^f+nX.چ燃S vcE Y` I! K>ϭWCb0Bd%dtvR8W< cZ;2t{ϥ7!@!'f#y_UF-.&wuXz[HT;e3q*ؙKL\ȊpH3k\͢Bl:C|摇=.=6*ay)I,rHf^A2zB堿6mTu',ZJUpk70f=>[I@ Ъoj8;HLA4Pv-V|'|Op*ltv?FݮA6ǷwSRwMzo5(W}?YSux"fv@e@B]Nr  8L \'ϻ v`@o%V![A;x+hi|J|譤X!%f7rH4e2jiC,u;] ulOXb)V%+Ph˰,- !5kwi( L ǵ$E\jc v+DFA}lIZjBrjH5йm-/_wRXJ4~{z 7q}zBw_Kl`C-"PԒT{sj̟vrrUNo.TWNdg, m6:(Y8 'I"M!|HPiJnhDSj;+H:8$c F!>$6o&$iG>Ea@$6Cr&9Tojw2/)Y z9Gl *P*OȐ^R2:ВNZuRYnD0g4aW*9SMC(,O OG3;fߘX82(݉~5ձL"ӀvqUA8„&|xp1a le0o KغCd8Yď!{:ۂa |T2KE1-g~_,w<5mb[7$Id9d5j~(m Eu$./_QsX .NC"{< N~L>v?|Dߋfr?/ %'O hJ$wKOX:yp !~ek-B^V:A.$~mRH|Pj'0!w?!N!dJI)8iJT[jAi]f_C,E2У4סK`dgT͈p CqlN[.?oAx6O'mf\fl= iհbߨ)br#Xƭ(P1DBJ2&k=њ1&Dկh pJB8wm'mnT[;H}ֻ;8Lƪ8LJ8jt7la'H(xHF(w{Fu@L`:ݳfr5WW!vT|Z+,\C MS\C9R6 | _f3UKg X2oku)A ܅B [/a >}wg" {(J7TtXoN<A\l7a [Mآo[t>G0#,Fz zNG6bI֡zB_4S^"N݉Nԋ źW&Մ{ zVHɧwO'KW_Mn4)WǚӹB4ʅ6V! &8h7A MA+Çxϑ^Xq,[D KF,W\`qsL$\Kcv!kߝ~ eRS<҈ T @ELUC昄ba%*<Ŕi*&seQN'w2C`ZοW+)G>, @0TɚLhVX Wʑޫk-\bXNYݵϩ. JJ L}PS"; QPܑk ̈Ζ9w໰Y/3ݽVA tTFH|6{ц " 5(q/ B)}?|ԛm$\},&Mϛ7E,^}P騊&;|9{b.Yr+n .ڔ;Q.o՗`$CYRkbq CJ⑀b MT{4bG_\فem(J|4}װ9,@z\kc}a_?Cn2+("Rr+5 :ZԴGt R{AbcHUl5@i*,S$$6>v1pS&ڋ g`;&L|nQqU8B{Ø 2'3BRHc`䀰vcZRxIRH)yn<g-OHBzD Ji-&Zicu933 xÌ׎Q!|` )wIR 0itb] Ml1UB;)l $@ Ă^BRcjb $ZeGn::q {a$qbHb-'R8g! AUB5yԶ΋aX_)хVR lZ EnAA(}N˜>%BQd U%΋Pڵ& h$$ s#w @L@S?тcӢA>+ն:աEރfP@S2(?'B#8]k [U3QO X=@QɭG4l!0DX~&&4T')VRSzۉir1YT0t.x4݂*׍4.8fo- ߒpn=5 g5р<-D@ni V`` 8Eo.fi*^uaݙolძj"8W[dE/r}3fvQŰdEXt^꡷ B||IYOs%صt#~W; S'>J(e1b7!QV11ǥVxLw\RJ:JRfט{P&%"G욜ڙRnGK]vjRص~RJ U{ƷOWFǣ@9FRW2mT1hIGvQuwUr8x3 q^?^hÔ*%ŪӘB:9IJ/N| zc Lbbk0% vLQD F-”(lk1e0Uh.)0܋h> 'Rtu%f0}U#fvlؗ`RL'?2&`}A߃n5B+*2ǔ{C$rY ݒ?jd@dr+!SzK?GF v==4l%9_iV.33Jn>&w#)+6Q?yT@-P<ԂO ҷhoća__F&($^,LvJOY[ڌBKe"K-_ϲ?oPp]әkk{);z|giwD[+m2<۶րL!p''DO?>{O[w2{ٯ|'LO^]?m[OHo7!th^!i4}?A.S7{۽s?3|LKsؤwB<ם5#kn 8H$!S`5;YphwttKDJNˤ"Yf%J3Xz ^8ysAf'~xw\|t; t8g6D4:q ":v$P LskYtC(?L<=|&fp)| _/ [hJ@OU\["c`KgYdICCv1w%\ߛ@v,[Hv}ѯZHՅ- ߢvffHx5Ϧ| &# r03~n6ifF׃0h Q6w}:Lse\ (͂{$pΏ( UQUPΓd8IDP$HکI:̀]ujz9R2YRuP`qM73.ωK-7cl8(مd7+Ϛg9YK͹u'{h6ݔ'Fgf,V"-8$0-V2-8(-g< y4UK`|p>+"j4BtSoTzSF "#EHljCCXDϝX%6:CS&myA:MF|M؝d.}'0J ~|OEhM`% }'u߅+|9>4f"v*UTK4idhLs$6R`6+SY72 :_pֶ-Wb ux8*$A)V!#5 =fkZ"*^ D*~^*X0[|R,`mBūAJP)RiCl uxbD|_: 1P4$FY`xQh_U8۪ƕֵ+[JϵfVA1ݦZFtVL)M%Ɛh⬤q:' ^ nا!(дT\餪I"Ԫ&]*y.sIVJZu{ҪkXdU׮)jD[oa2Ni.Lb\` V.f)`&V0fm遶@[z-=p~k2M T*KUw PK#+lZ/א))=}LI( ' H"T{% C%hM0'Uc%w&>O(:[X_D*(׸<i$Gʟ^d֖ڠٵe./ܤeYiCIyJ呣Yy IT:bǤ0σos Q\" #8V]LԩDrOBZK,} ZM[deL)Vpa0R|ك{>vS'es:rH{ӛ7o.«^L¾jC~M>m(u=r>I|7ޕ,"}CÀ1X$fI.ݷAIīi{ 0IUuw]]U=I=`]~_?`?;n~}?էaFaognxA)?؃Զ ̽"<ߧof/KbP܋$5wr`o%W>8 6'VΫ|7yX1 ; ]љ j5^d)p ~p61sKԘ[:3e~xtxcd٨UJi]S?<-GV!XzX!`B)3o!A_MccYW5IAvoZ"<lZ4g9&f%}(1TG]ɑE^d}*ꭥd?n( t(]%Q5mD~[sR-pK_Z= ~$S߄! |K(4`ɕ3`!Qa,2i)HXaL6\EJuP~lfKsx,)Fg,< mEE*Ӏ04ʏ€BA`"#|_Pj\ѐ|)j@xp,Ξߕv UYcu׉} 1,9mWjT%c?tU a:sy!`KRj|@`v}&`ē/lJ1g]+߼$+S`dC M뱰jkH±RD$ ʞ ,s1)0ހ}a&^8Q(1 `2fQ(|dpXJHo+,cŽ쫶_&šZE a fTEHjM @ɐROd Yʬ v[\/B D@:x(D\*N \+d>^s+B7ʮ藈 {Fs<|HD#* %J (a);_G~T҉/58s!|"N.ݨ/Gg{k`"ٔ;u3L@@Qq)&%8/dPG1?\"7cd}qFE0|ЗQ %CNWXEcĻTfgd}s𷕓p=-YrFa 6"!*cfr%Wy=|ET}뙬 81!f6D+TP48B@vP: b " E !ES *V kh#%Voj &~._ moMza}A KrTl)X0e0[}6V ejFl4Im&w&)S[` 9R,FD)pO$/e>jzϰfWЀ(*0A' ԅ}"OW"mL(mspQӆY_Q>l'%@cc/(F3ΰTs |+j@\ gC7zj ee'z(GEhӅ:M.>)`/NM+?ŋhwJttnxF>ؚ R~$;42L2/wf7^x;-֔|^je_ɐ‹` xE]~rlz&0)S=weM%iuwVѕ׃.9{qgyv9,Lt\,Mo$̂(DJgW\_gS ٻ~iBHX%DyΠ-֘Tϥ)ygɥN@޲[Ijk/ǻ'$ ,Sf3&[љ=#*@Aq3r&H^K Z=[we;k;W@{B 3/߰8(@rŎagV~{?OK7|si{8'ũ R)y০v}TxZ. vOǔ?7$&| Ao`wV+NXi@5MNNv9dOӒOo$HVr@n4˭᭏W+wR~, *~'f,9su[fs NrGZ<ˎVi"+n7 V\[ dA(#_N[H0IvaZd* 㖌\l/ebvbkg)J=q qc҅$-1@ L8nvvXjDU!oj>lצJ#7c5p4uYk îG%Z; -iӁ 'HO8ܠ=t@Ӻ;+iת>v[3oטI2e VҘifcT`Z˘ ߗߞx!8aC[_aZg¹,;H'!tޕ]_,H):Cbwh0h@ w_+(ph^}ϩ5ץf~ރsB܆V P dL cǹ< :d5̰z' QJq2P FU0<=~H7X +b|bϨJ4&$' –8?Wh0mNyui@V AX) -8-i8'7S|]tE" ݻ]fY^ \/'Sx˽Y?PB#G\>lRbۭO IloðI: J0"Dۿ4@m }5Sީ?94ߓ;J}.:!dlSlZ D[Z6NaMZGIe8]ұ%Dr>Nn3CTJ*RRgټ (Lcy-" |U c7J'1&~+ oB.܏""HI ,U 8 `1o5 !)׃A>kX!ػ&i`}F]cDboEOOkFUZk;=ybMЀj=jh3tM^x"oO s{k5E.w6W+pڥ;KyS%As U'P2.7NM-aڄ͵$ ݨ= Tvn}uXJ_rBVr`c=?-=GQ!Ia8H$qmV}vzmDe8_ T 4 K"̱jr*ܳR)h~yhOg8H#ٯrZkـ*cM>ljs>fW`غhD d2SNA3¾>8xBQŏds1xAˆE~SeQ~c}c֚s4ޟXKG5kwWD4/8;v?c0qǒ_y{RAr\ ٷiJtjُ]Ld)42DZCU8H87k6ye6¨ :\FR(15JEW9dy3V*B4R QMlNԈ[N| rUs:^}?jj2GҦfNohQSlaFNeЪea[7'$9Z5Ge;3@f />S P9Q/\){/=*TJhtikՀaqv+ V#g+0X&Z,ځ [7m 2lYsT^hIJbd|C'NH]@cG[(Q#Hk*\e>KOc)M ~"SԺa!yP-TFngJncWgGFFPB#N~5P+4&g/Wd/IR4xtn0>xN.:[?>eb);(xY)O^jtߺPĴh$Yr_.m@QYN׃|;f(34Q9XsjV2uv$ x<>\df㴿)BJ.޿Tn͕Lu BIInZ ȘL5if,.yY%ESo9 g%Ƀ٣(|Pw1XK˅͖|] 8 D^ Iv(( ',)  HP.S'0c/EO]J'lًc}N~3ink-߯ SC㟿ZABc1^X>J1#&KzCz,@bggLTe.H 'gt'RH" ޕ26`H}4ժ>db ɒJjn)!6JT'+DB܏=V^fԧAQ۩QB )Rl暇c3x-cM H*jcߴ'p4A 8ۻX6AabkQ V=sD0B5 -j!Ԁ>m)zYߧa؅GNu*`m^XR}i`H+>nr˙.6&ݵ+!Ϣr#rY:va<4 ~ߍGl&; y](Jv۷ RU%PHჿIX{KOۣgT:/?Qs@*8 =ȿ٧GGrQho|\cCȻ'`iroBDxzaNFf<'4_]oW9]8HrA_\k%ג4p%YZ\qw'8ejw7 9 +lhv&! 5RJ\{pX \療aF!A8,3)%\>I& Ӽ !$_%ʩn~%(g `5QX铦xq+č vn]QI XTKe-uCUFR-QJQ/tjB"$z?3p{kL.=*Sx wQHh'ybGq(U{>%b$ skdv4T^'ЫԺ㛆)2.8Vy ILr }7`n`r-9H+n)ҎES %hODm?m:7=9/H'7Tn;%g)ԌP}bc߇a= GTEr3*z׵xá=glpg-Nnb/m!pD `eEgǓ/zR3zOnFNȆL`sו'WiN'DOګyóII!n%}=<Ǧ#.paR:x)bxf'ɖlFptOxFF)vVm͙nrU!_;:|>P5S34OiiD9EZg]5\0:zJzQߠ'4ޑͲ8. _Ҁ^" sҍM\v=5n{4(C]A]·ZjѸ)h.߲dT͈WRE=O (=[_]9 ":5LMBqyQco(qsriQ6T#x^!}!iDU0ʫSزМz7Il],~SW! /0nLqZ|iҍGa/hfBlӦXv5A(1}}>]ܹGoL'~K,1ͻ<~`vDyU;jjshRœ\균 }.uMktpj⣷ꉋ^ dǃ r𭹻HdҒӌ!\1i/r*gUNt$W1o =;Kn fڰs7WvHuo$X/_dccI.*D8m= mt'6?6ILbcWM1=.&IGb{-ɈJGA5eE#*2~`ݳLAB-l`8śW?'t~ ɢ|BJ2] yxWw++oތG5,^|=X<ۑ͌`dxxw:E<,}# {|ݺ9>O(Y5z3 o=%9MD#z.YGRiŹ>\\D!\uتZԘ[(Ax|6й;)Z$Щ%ɖl73%´Ry{M5;{;otKLڞ}ZL" Cmq-*uDg/U,/\_j2jسln•D6)GQsL>\'P?WNOi$tH9&XonagZߗco&030yjͬjGcˤL4uw^ʹ@L5Ϟ$!٦ R@њ'%AQgA9B1 3g_,Pi)6\\e5.f|S#[q:?8ZkŶ9wm)-+3+ɬ%V|G.d<%GTZw:O7Did4Nk: ``! K [ObnLb~yȫ&E|F{A}dR'JJ^=w%ȥ_.&5' ;/ \A(T#8-+[ h(߭(#K_T"fj1 x` @V?+xrlTh+J"RbW@!j+8j znzbt R/;O aA|?x\"v9 G<[@)A呛-QFKGۯ;%βL)Xtncxx 0aLƼY=w87Ů]_uZۭ rBs6j;dQ ,;Ya N1P.&s*MNѰ+JMQo O5z5SMjyX\gN(cX ]ј`fO85!y|S9 J\sL@JI+"Eؒ6AtM -ujא1ZiXVpz;FZ a#Csݍ(Ch CTuA Vhfx(J?JgD ?C > %̸j98.>Wr zQu$k3{$ %QzWmrNp%YƔ41Q)gu:sL9ÂK l vxDQƒ=\qRϫV >DW^o-pA ~^kXH{"AS<ʑkrӳ#4$h9s<;X0mh}tx>x {YvL98"Q)2aIQz߳JewSEpl:>!2J5H678)3YxFV_12!`QgCڽT_Of;S0ݞMdk2]vZ9<˪YT=PƳ%lERa#Yp/4fDʺC4 jڔ̡;uC !'Ω72[;|+Ϳǐcڹ)0L@KNhC8VuFY1jT4SԈr%Ou,P5/Y)Y\8k@DX~ƙ9Q l;ZC{ռc< ) ray" sz<s0w};O?kiwÜG~2/(yc-C,=.KD +=~Sf9>\nBBѽKq?-j"д~s"!(gfIָbr[zm-^ʪwYMxp(8YCCp"r>DX |vG6+[`fnƀ]s19⊁y_u~J :ZF1HKKz&btROF[W ;&Ůxa,{yO4^;FGߝ"ָ^B ߭=mA̓˛%Ub;t JU9JѢu h:բ:T(St(tj8spEIDyQx(-8['r. T\ьj˔yAĖN rŊiJΩG&<>O V׺X͆&Bk|~G1iQs̷iO${;jORcyYC95G URf5m.Q<"|{G[ޥ#˺s1DDsQeFCFToxTܾwW3Љ) X3&𐽇[Gtkk”rIr\hF(%J}7yOcɖ8_6f{4S>EVۤ@.^Cŵ{D!jMBl Vh^>P7`l, byz"ADo ֡}y{o(D`}q{ip1l^nsݾN%r0J o]GߺLob `uߙeF;lfSBܟWൢ^KR+PQSDR"p+cq.'Dp$I*=-C3W" |@7{6 }ꥥr47X&\P)v!w!P)QhM"%G2ϮTK5 . ϸ Pks) #cM^*p(ՐʰhpJi6l}zHBu+z+ɺW*oTv؀R0SxkS+c9"6ړ4rIig/t`T6TzJeTM:=Y{ 90 $Z3!'lC $PIR i zs"6읓4t)Lu讝{:46䵗mܔ_Eq̢T?YT5bO3ݴ} j)]O#]O4nWo1x0cx~2tD9#G&NrIF<*$ ~ͻy_u2x; {E"<(]dn!:hͮ'EWCFY`v9 Fx{\f#jv0 aæmrsPB 'Hʢ$F1D% 3HeX;ߚK*ôfs;4˞gUQDu?~**\p] &GCahGVVTN=`f ΢}7}:OF3.9fBq1Z1_/{;zԨTF[;K ZDڢIHpINyё X$`B IT|e4{}))AOFBG6{s16*!$q+|=PF< Y8A;Ia*&)7*$H{*^VJMbq(Ozs~2vTmSEQ$p˲NV);[ы$ߝ(ZC!Њ5~ȴ2w  q!wA_/()lNޖ ~ j<,EP TzяzK*@eSYE$!{Pd"1%{J Ƥ1=G)z(E7O1ާd@v$`W-oT%amVJ URJDcMFU+Xmmim9EH(^Zgb|iLH߈M+>ή|4E?`!mLPQHH]٨]=A:[<o11̅9O񢽶njϔ^o^ x<_T]]մ y;yݾ3%zU;5_<$tmQ4d+xSDBN̴cbgJ= ;8{c3Vۼ<|n (lH7%zs8p}aa~ :" }%}"Kr_lN}XE7Tl)(#Uy88118N$DוP:I.:JTIs>P+""xj:&!SJF ԽYn׌Au\ʚcv=yKyKR{]xP3"CkkR; (T3i@%J] *R,8gbvae)toY69$&+lB(eS4X#_er0|aR@TB.jlkV.]Kߜ5 7)kQ0[kba&ٚ6&[+KGr K\zyQ\!QrΨ@A`rcl]P[.%[c$ANՉ.:#ZQPjBT'jt̺(h3xWq5 Z?(8Iŀ$h˒U+ S!$!i- +EgqMG|Ԣ+*ѡD$#B'KLIA$UC醏8 Ac\T".jcUa,Pgg T|JS:iJ7WP'.C[fK%d35e>K "|rJ J>QuA$Z49a?O: ԧjWiBgQ,=%h [SÒZRr]9 jH1h<)Yp4{aQ l,J"I QBry+gqF骢t,A61In"WlCCT$,M/12o:̹R8;GPN͟뛛b_nyZéj5֯x:3yDϯ<yem ʕt6'#zXhW!'wv:i.~r__\eQҀq4Q&;oѸ:s[CyKfx ?]]Ɵ 5[JM0#so)BYEx-IC,SEtD]#W\Y$!8Yg狃ݱgFsڛ%ͨ[fHVWEXU9[& t 82X4es 4.q9$#ZcMI 5 422wݶg -3pk\KۿioRFolL 0:-(9eQYI?8:TXr$|g (.ć$kt [ݪ'@JԪ^b*bW=|D9]%PIIՓ#UB @py)n C MizOxDu6զLcd'"/>F&"*Ҭoo^y77F]5Ad ;}3vwn "*n.<6_7?ƣnQGxjT^ix_<\ 9$r9W|X92ݢgʅwv8J,oxpk۸m/=ܹ@@} -wCn"/jIT‘}ҍMy{.p'?ܯ̓ja]^(' G4zB k+CgVG F 7u{vq…BX{`i=i>PAA%Pbsb۾\D~&7ƌH”Ҽ{Xv3AjpP=z)$_@`1:s%Ĩ@".Z?ySz+%IWO^9N˻hm ɐe1{&(VlDm0Sv30 ɫ3Oш%˵ԸU&`g?nC5|LR;2$54RR F[o`HޙN#YG*4@ #pdNcYҭ͚P6Lȼ !Vi=W&q p3 =x (PP8Śx%y3QtRCw(륗4'g?({:x47K 3|$R0 8<f3__.NQH$6(ʅѫKID$2_ݎ}Xe<5}u~WU HVГ$uqy''V޼֭F%ҋ_fk1!B(1ftoBTɗ)ǒެӇ4cWaE8>5<{>Ly٦*=f. $/ ABP^@V2"k!'f\^\R2DRT$QNIm}:*K5rGc>`sU:w999xm Wl#f̒Dp[BʡrԨYΏ+)(y÷RP,9!-eR$UT2 }-O`͐݃!ל=7#ݞtpٙCkJ;N}bM+,k`߲ҶUwOwF8;=6ye rӷLA\RPٟLr`t]b \vwC` : `1zp1&*FU,$T+_ .zRbLr NlV]{dt1S'&a6 .Ax;vF? ř³fU"Vrӷ[{ϵb1tp9>!H$d< F5~lf9RبOX4a9~Wi睽!1d0H{B8i_o?wRt0&}~L]}3$a5J@Zq_xkXk0_ alծtdj蝉8xǼҩr$;'K$k ZE&Ⲹ)X~n&]Ov ^@Y$N^o7fʬɇΪ@fL|J~LW;*m2G?õ_NHFv[c,пYhGJ~_s?;ɕ[|}>s㊯u sWqwr?YET4w-tJRN% @iio~&\eƸPU;~kwH`|}MwNr(E&WX"a\,߽n*cC7Zi,¥t3,L| e7*T-Ň_|ytGn1K  ^.X)Jz,JBh+₦_z+{=r\!f;_1I,ʉZT캰Gv9ѬTSG)0WXM!hEi 5K)J˭ڊOS/֧he!2e! ϫ{YsȆ:ɐE7W_ywS~E.~E?OMj9%O.:KfZuJ ]Hι^E],oJ}PcՇۛp;mE%j_g0G\TOrRJ sni n"j6Hǵ*T\㞲Hp@;t!ug[=wPԤLGI&.dV8;vYbPs͏uQ-$usCYԊpEy2mA5*Y+wo2tnlT"YrC^)6\O-NXN矹 .e,YAc^P %.l7^dғKO^DR}T7M2:LF[cE&ξ۝wMWm2OƼet쟣oC%=i'Yu9## ~C׳7sc[mLI& ͈2uT?g4-h7!TEkQQÌ䒊KǑt DiE)Y#>,G8Y)nu饱(ұnvWt1qd^uezHk87gǀU9M 5Kv>+ |G+̀O]Y,[Ys+KdtR HFAQ'KTWdZd#ƒ˭s˥ P5w!vwI̹sC5^t>s̹,tF~FNj!Z9Er'CZw'quIݭgg 5V86K"_,3=Zb8 rCԫ:Q'rl8e/Wi O"HFft#'JyV9z!7RIt'Qs[Pei]0vYH-pjA؁ JQi뉶"D˥ρ3:YJfldIuh0qsYLa,q% Ѓ(.)f+FtIL=83ka%G[#@yKZSFqc*d47M)Op֎PvZS먃l]>7̳:8iw J`@Jr#`'4%T+sԘpv` Bs:iaK\PfR%M~eқϏ>~ nz5AҔD O$XX }8\Xe%:B;'<(*n.K:V|^AkѲ bV?)U볏a3`x[|C?hB1Di6̱jEף 9kVN4-[#CU5߼M2 Д1H)u֗^RNp8 *l"HԶtQ)c]9¼LUZPsSx&|̍~]W泂;'fY4kŞbvNb2',oi\e [ENX3FW9aN9ai 3+v-F Bh9akÜ%vϊ@\t}a &r=l83w:`Isaj)RCJT<. 2] hb;|_'LL)Ĝ:5M9l r5|tN941 .qR*ƓHT'{,1A;NX~gvCjMaǠZV∭{Sr!?ܫWNVO7|l# #Xw9xGZ6KzђW(DXBRR n|ˇ>%H;i'1Ph 'ܴhu,#ςbx5Ȓt;΢_@ g9Slև ,IWyNՊnjpX j!;y %=mU=X흎y.$ˍs QRJ69uXH Lp_rDIUPF% U(0@aaC$tyHԝ`7Ax!bKk|XVA:Х"2J!D/ F%u9N*:*JV`גkfGqI"́Z6G2$P X%$<3A`֘o*DZG l`if  `F&q g䠈Lq*?l vaҕ5ʖơ͹VgVg~9HE{+$4wڀ;7?ztNPuڊ埗%yD?kp]?1& e#b"MF']I$Ħ]Zj8u ǸҜnѳYԤEcȠLj NFokJ`0gǑHkYmkŜ/ΌY 9V ﰑmZ(.|Lsnn6sѪ=W!J|@=^S* g3 JSɘ"`Kέ4'AhS4*=a $"0Bz'm7lK_଄ NRjÀÜ.C%+y&x8뺚m=¦(ywmr[ Mvvtw c$Bmɑd曆$cɲfL Hly<$CC6KX)s[9g a$ܕi,gMFbl S'q*}"L@[R`DeabIB*$x/<:2䙙M{7n1;)YY]%sXK]OAȍłc~勏ȑcTp #5s-->GNQ9 P ;#wQ2s=^|;*AW4187`jG8%)l`Ph= -4f\p2GTY'\y]`AtF[G[K`<'AYe݅'rUb ?rӼZoi[ѯ ^߭MWVɯ(1&,̦ DӜWr2w}/| =+rk9M #rK()A)% :xeZ]@ %JގH-Nxcb"ޮzwz]zo 5sCyuh(˽)en4R`Gdh%IݫSbQ`LY%/t bY>5FbޘNoS e$|m~T/_>7^#m r]RI,fNa}& 2c.ӻ $BVwAtgntKC!+Pwe̞v7߽wbҝaxpJ~lmB[[4Tvn!o4̢; tU[FAH>,gR;  :ԸI^m*7x-RQsGA~CV ݹkӡqAìDlnCX|m4=h1Vw.dՃN;Y`Ahya[^(lZ;؎O{ý͵2mj vHq55I أD~ALYQWh({N^@h㽕Q띾DQ>@iV>DQDQ&_zw*_D !:5x.S?S(uGzgv¨jdYzcHŶ815DR^8=6914?٦Q?;sϛÜCչ|6?rC4l=m ] uCCQhmXڏC^ {bϞGY1Q8lj$ qANx}8pzgT7!.}w_FI^x:FnF\R ny3Q`Q&"Zp\~O~}P~8訬2wW7<-0zq]eɳ=Ovަs{w}텶O"Ve'n& ?ޗOzVO7_CUeO51+ځGו/jOӕ[E `Cؕrl b{q DG'`XF! !E7w*@*j=N }>% *1b bj>Vlև&e8T :}qkĈJyr+"|nHg6E}x5!9ޖuڋ%:Fi ^,VY F_2 Nq 5`bs Q!DSvV7U kQ *MAuh5g桨4EۢݗކZ~qGᖳ7Zİ2s$慚"#S&ؔ$\Qh 2AůY?&WTJߣdRWjcg?sgVp{-1f$fqԦdh-b^*:J6yfj\CsPCmp#+"hQz[^kqj1pzӲ}-f԰%'o[v?†ݏ|NQC"8(8J>j\uPshDzfE9P|3+ oSmJd]ׇ5[,>>l('-KtHQrU{M2l?Y^Yi}K]f^eMY[3-ݚ1J^w|1OHty}d b|R~/!C̸R ?dF+:CppĺSd y羠NPo). HQ~.% oT{>S Nk\N/J(Ewt!X'~zVZsrjר53!x8gfsڝ kw Cl['7*9Y. vt a@y!S@ڴI:nxI01ڜ^hyVT詍!^ KGUh$@vE@֤̔ MR!O¡I!',R!Gbi 췳lU ͖+!Ẕ^I]\q ^,_g|_\թpw%z_]xT-Cӕ*S6ll-/n??HcacZ\[̅F6s2GTY'ΘQYguD 鰵b~!4ֱ lF\&Hc#Ld2 4h%D!Ah04˨)ͤ@`%V f,9έP̙B\'mA9W08Kٯ=CU=iQ$Q  Rpi21?2k)QT"Y LQHh_F@La'Ȱ4Ί0jPi> R21XPCs,A@%fb Z)DS[҂$p@ Fь*PT3{@ A ZXS0<2,ǹA(60rCr^d6G`+`F}@8YϝTeX`.wYNiƵӂ"n$M )@>-HhVv *e_bfg[X vEjy@ l/u~"SZ?o ?]90N=#2R|*/DHퟯO槿i+eeb7ww?ú8zٕ-%`-mju6vKέ |=pg8BzI?voo܅΍BUVfҁ} ݠ@HFL`,,֐isif5)3OK+qZzSTP̘c+iıGUBճ=W c1=GQ3 a\}C;WP# #N3 a\}C8;WPz }dJ}jt{IPQA4W:`0Zp6c8Aa\Jr ?pd\t!&l}CL8j)o>[3q!l}j 93VviOꌰ0Q`V[iel4Xf+%ՔZCäJ ! +v yޫfRV&F^5WIޜ~vջk5}?+o+|h;i^Z|+Gw־נߤeW[{d5QI{-%G^{pZ,ܲkֽƔJsWzנ %媷Q֖(ny~fܤ,//Z.f[t\xY7y74ĺN]A|vﷇK?to۞z[.j,VxJ$5}`Ug+=N'Pڊ_ 9s69ĺ1QI+h|1#:m4n=r!fz5[ 9s)EȺi`|1#:m4n=rz aԛu״n!$EKցuSڈ:FoDzTެRNMBB\D}dJa=]}Ƹfi<1%:m4n"q?u !g.^2eM<sZm4w-X[w-z~TiL2Šĺ˔(97:_Ӻ3Q/jU1fߺq/uDǷ&-^FUɩ/f8u !!g.2E%w`݄䃈bPGt|hb"f^rH1So5[ 9s)Nͻk+j2?Sg@M#f B#,%R{]+eyI赔2e죠0\xS6)eL&2%ucQVLY )!&0Mj` )"!5JV CR4EKCj{#JYK֎Yӊ"kYjbYӊYK]d NtZPMP-F\ZӄRVY^KO ŋDMY E֠״Ǔ:)vR[d &UZPM` ̤aRd "k S EƄZDd+ 7W(ksw;Y,o?o|fSI,I;%S٩f:0#b~~w|;qn^Ұ,Ǻ Y.%gkos @q 1/dNslQ\Qh8hN97@-Og|OkKN9BwTlmβ,Fc S-UfXzbp \/~{4~oMf\9lrPNBYWB4ٻ6$W}wf0[3uúAiM1YD,% ÖxT"- m,f,43G`BpNa"c$ȁG( *8 ̣85.ЄU'kBp2')D J`K4)("1LXPg1-3;/`cc/DhG/A%KL ǯ)vτ MߠC*T -_Ho4zRI$1Gl؃FBjhb)kBIP>Ȓoޜf˩NŰTXh:7).8/go8o yU@r*pTHP'7z랷㞋 B=/8Gm}w(+[]L..GTi~K*'wbJʫK, o~:o$sXN&m\繆oq]-3?0ԧ3)"׆l6/}L> B8Ѕ4A,Syb TBa0w_?_%+WXʡ5 D!owd-tXEFl'c?O'%Dg_(5N[ff<5Zt,_bv1{rAދw.+Hen@d&*DPo5y@w/{IXir^B{q4ieKO~8zpw<*y2C%Q7/6X8;Xg?)_׽v+aJh۱˃b.]YMKb)LN\Yto ;f!;ƌ}_qwTyg 4޿ߋW(h z0>^\& 8`r"7 څ}7u+MPcs4LH%bާ.pÅ% $p*0IZR.VH {bĻOc*vLw*sƍ(}p!5 xIE٬T}wٍRj qZ.v2i8ջ܅6|Dn7g(#uldrBuJqxMŁu{ācQ:{E<wۋz[ı9NSɅ>#~C:vY[]m;ivLmxF^K*-ku:uFykp➃&,Hy,2^˒K"9)H0ݴu}Ek14Qp%=0csi+[4o'[| Z׽NU xw]Aw]\W$#Z5륋2]5Qt&XKEPTh*Ye'<4PF=5h+hcǧ$r7 jɍJc;%kH mo?36aaDe6ͷ;Hm~bH䳫Oc+bfCPhJYG5J3EIЁhҎ{MSItkHk}{u[V֧p0.]}>S קX#42S]1M)bP"Fq8:sq&#:pZHЈ.t?=Ol[n{?uY ?[Ļ럺럞Q0+ɠ5\)θnQOא>O) Ԟ3]qZ0Š/Fp{z2J Z4PcMX)-grEDp.$De^<&{d2/鼉BPSaWՆo(4gJn M s7ZJ <"UT)c)6j?fjMͭ)l[3:m#,w|'C盍|#8Pa8Q)O۫,WS`Ny9C1$ I5xAtx˸Qh1>i$Qe><4k++%p(UVR7_!(#87JAJ3@;$'R%II:Ʊ+( 8jlλ85 dȂX*(TJBn|(QtwsCf7`4޲photJ#U~/ڭKL {w]9yta8 x߇W;OnQ{Uu+!.F|- ɸ56&v_z81z3h84u-!*kk2N^/ivCIfwIc)'[ܝ>Q{9`5Zbt -B4ӯXX".jU*Gюcs0,MQJ$KkmLT$fӘ _~Н.>68ewej*o҂k.¯~*J̈́!m>:`oeV_k܅d CeP̫D9G/u4D*Z{ jZ0.40*6dBmRmME #.:R>h;_k ×^K|lѶo_Ot͌6]9] ~l@EA⚓%a߳_&i/I +`Yl4]N1l(>F^U*?\g U~PQ|2 LGetq&&zKQJ4uZi `zڼ-b3qCL|Դ徺 7i{gsӼl4Z ƍoequi3 qA>4g{8(%ƅK,(Th"Ф>FGӆi'>\6_2 eA6;fZm vQJ2l `3R d")\~/*T s>؛6MR2ޚ9ei[s*טSijs*X$)aːB":J{C7ӄG(h NРB#["a_6>paӦtM ;}ˢwװhS qעTճ(U}[vf>ђ+z1fqPK'ƀ2BL */h-Xe#"<_J{XǦZ1f8ZB703mG2ʦBӤ1ʂT&*JFaYHmf!Sς) hƵC4aѦ|\%c"\0A2"TlCm>C%5b)q $;q0!2b `W✀,4 '\mwL][o+vb/;3X,0 `8'[NNoQԊ0dFdW.V&4<6$ DV&AiBX?Şʺ;>'peФk$5D3M{9XU*gNbpE!Ok>cM)T\%aj\8,ՍBq(5>YBM:ڪmdT0u](D)4./ kMbZB6_W6B&orH+2L\Ȳcc6I XIIԁLiA2@[m74i۵&e!b4s:,BMBZv! )ȻHghȡyX g(v1-Lۮ5Yz,4ZO7Lfn@Ɩ[Tɡ@MmCSBƅ4:R93bm+aDM, o\=\ h+&qh4ԍ}JKҙ֠lhBYǨ׉\ Cc:ZbLC;oG"k_gܸ݅5]|N 2;˟nw:g CBN07_GᗿiyY?\_]V SkwsX}&'u1ZGNNcj՗! b]# N ﱀ]H;q)ztu~xwA+c]qkL3 jc%o$eR|;L(ekt}U~;qd |]y#sיˁQffnGNI8-ߎS5V=;4/_}mih>~/gW/rض5LUtws]M՚t;Uq呥ί棌ZUhAVQT%4%],^ykpKY8pa&M7٬ٙ ӏ:O(=ui? ZXtƞAQ˂azazazazކi,(eSl\#kss9ԁW+O.5 ʻa߾,x}V C9ejMH''VHo6tAnl&P%c r-ѹE3e` aro_p|jal(/q'OVnjUX.?ݻ g&O| 0?X E'LG17hD4sN5H $p =Hsz#BEwC2ZH@YڏlG$%̂GPڏ BivoE%yp #Ivd[n㧒 ՋUԫvC8|&C;֎԰V!=lyÏ>Q?) 5FLػz΁Ϥ/>{>:[GHZ~ҷ9 J+ &W!q5kBW;ۡ[eFJ!Y;aʪna E4s+=pN).kwK$rOD^LP.U5_$̀Ek3V ~~ߑ |MXoǛP+tՎglj5?KQGr2EK&YZw8s gRWWE&YmTS I厵S"ƌ 0ؘ5} {~4_^ oXd|x%md~/zfL>NiU(0H#U\>hgϱY6JЀ.R=zwoЈtSR~WH6o >BL_K>ʬB=_QW_+T`TXCXoW~f.bPڝ P2K)q#Dm]=}X^Ҁ=o^E'8Lq{7#8?("Sh}#~yptYuDrcP?_k^?"x[gV"mG6LY.?yddNLSdHvNѣ{Ӈ()}\ޓ3ES=VLl-~ q41C^ !k #D8AѫA)}h92k f8?Z\0q9ňиqq z0S'YRnVV~S]ƭS-csy36y~=s_Ϝ33aAxQZ^>8[ !bC%C: Mذ46l[M P3 aPU5َg;vycm;֖ʆZmDV*tLch;D6%5`]} 9/je~䳽@^xVՔ <J́g(Htuz!s]DOSAϝ>:%KQ^ ƌ~52]3kڂFxr:J_['_u޲I4:Da҂ykL5`eˆ4G "1ڛZKӑoWxNȒEҰ#BV ¢ }mj yP$!3B BXopB()o0Q[ .c9EQ*uumPD&hIj@ SOӮY;#g`vl}|&+կ!ޭ~s`枟)#,Q,Rk,dX nF+>:NmCP 2&ezvJ0>d/.ח%6 +5hQflPjKO a|BNOV?es1i@ޟ*oɟz&JX@tMZ9K6.ASw`-%jvqc}{džȐDZdÎPQ(GidI9m=8!^'B0;(ld2C^*VeUL4p-V `L.AlDag@n{ح=l |h娈 Q8{?c.n?A> ĶI7v];—nzfw$b8 O{q/\^|ҭf*g}:QT^=l*g> 6Ƙ~,`n8VeBFuИH'%&E*pp`oz?B:`,B(PΊ9N\t[-,(EVK43&7[H$l v.jYnx9%吗0tzb%Yj^2kP>fܥO/wggXg;2ܞ]u o.5eXꦩ"VU yȲUQFU~L-\ XZo *WObx6-􅓹iOtpɜxYФFjR}>e4DcH?4;5yh5@Z56h%+B{ <'5J]o.%J%7{bqsDz,]]8in:pg15]<.4W_.m>,צ+ztuZŏwgIݰ!LB[c vS% $eA;oR3g-َ֓\-_^vʯ ?{ܸ/=qˌJfg*}IGLR )٠@R򌒪dF tmTz*{":z֪b|U7-#-~5EeOWLڗfww_8_ךK2 25] ʜ(ݹ)Bfp3n Q@t`LЫ$>&I;XVRmWo}_NF\;aΰ.78O-Td0DdL GiI$|tYٵE.1hEq-Fi1V~`Pn] |Bf58lʩZLN#`\@Z53*e&7R '>fƙD1ZBx3~_1H D?c7"e2˜[m2 #Bے,aǬ8h<^ר:nPO>8--1ZTq} .^ 6%3n`uCz;qgǀywF00a4QG{Fho'=F6~C^;8j7aZur zO#ؒtgAX?}P>CR3߭0cR$0D*gFl}vCgoW2(%CNPFhG/]Qf;ܴS{Wd诃c3_)ognR6ÃKK9\"O]J=u6$;8AoFSGE}6]"D#|hjZQ66-;E(*{dwXMn)]xHQg"Edvߑ"8ܩOb<*jTcZg$ֿ0'x04e{o7<\hY4E'FO +[Rrc)"e,MpB`N47d,Z4 6rjZ*C8FHGjCj%rD=M=)A\l~^?MS{:~r%66ZZ|F)Ѕy6v#X4MS 9BϦX̦eBkV|+a+׊zE7jBpߦeG{k$&jEWB\χrnzF?V7Ԅ=QWõBji۵ Q(<_a(NRh[+/gu4, %A)Kh˾0*S\jbq;bS\ey6w UMOQ]jNR{,%M3k2A p,;w9m !)!56LJ6!)erHz(4iq d0OW`h[p9:jls> Z0cM -ΤҜFLIʈ1Y9R#rp3i3 R1y{\ZEv"aiƨQ"̺E\jH%4Xe%c#;!bm̸V<l((D<7 8-\:͹0?_"2pTOeL PoLg:>J_1At?dC4lpky+vq߁r!8:+Wnmy]D,dj:%QD*w_3M;+lSM C70v'z][.[UOѣzUh,>ꤙ}$FR:xI4kɭEi{6~nFS\Tx3g¯/˴Ū"7hDp"_BGx/&Oә;u"$LrH# 8_he-'ɡg'qm t%.܎RWtg 4+DblV֕ʋ(bv'|ISFdEq4I?Mo"fƨ}.+}g}.3i ᕃe.;' Y=ǻ" Ã{v>@G^ sSYB % iXfMv&_yf dmI i, dYS{:{>k>wiUxyTGhXݕ4:R!#xN?OR8G;MOO#I-iLHק IBaAT\Sɴ/lG^9嗢vS X__/_`W&U9^L 4(FzA${5J+*Z9Mk? > 78Z'"gSqR0ř!&8 bdƥVR&>Dtߟ \.WxXQl%~Š*[+WU9;^BlE7f$о 8ahX^M X"A]Om+Ha V Sbel2dBIT\a#uƬ4D hy |Hĵ~x_/< :~w&}ݨ<$m &7g,0M|c=qɹ>.~-䙧|VɊ\ųo Vg).DyHyb٧emN%_t+dOfrΙ '(AXDBrbnl}3i]fML A{\lg7_͓ e THHy. cC2sǭc'U+a$ BTwxX1"]*7:M:Cji"A8#J4`y N9HbU`MA,'ejUR=};*/mQD0Xx.yv2YthU8 ^K Hrg$J)+x( q=K _{C)P2V$|E!"w\5b,v;Ap_K:bΒL.(ZŶ"Z+P5iL0{ݱ1bO }1Ѱ;61F_M#3FQcd%SHqw`4 n_uͫ:5VcO=̒fW_?1^n͌d%k_g-1l9^-:yIJR/l3 Nn.aXub`VDvkSͣVXP EF9u3[}|[*렇.n캉8c˧{v1ZboQ0Wfd@ Nb8R!l^svz?gK)#U ӗ^mJjȵ\9!->t׊ NSxsn%K%q4:RSk񮍤3Ԍ0_Þ'ɸz1#+񗮤ݛ,$LFWR#aኑz*ֲH$7\M2ӻm`g_UN27M6+ͥsK4>XhbXm/nZ|Ntv< $KRI,,LMڧ?MqeL ;,kܸYMeDAKhM.i٘|l1nNlMf:S*{K}nmXnQ6ŏs9qz[,!FVЩwhֆ|&jS<0!`tcZs›ӫ[g<> Ɠwş__4 z0wgvdϫf-66NSXѾ՜t`km7dV'&$!)N ZlB^ 3 ɐr"d"1'L ,':g$Ŗqaw )9jW#ZS)$ II`Pjj71A]=yR̙ V|L3-9-(Y)tPYV2H2j f4#Ȇ)xO/1 \Hy͝!L(_ >OW)aScphS)F+O4J(ն^KF($a/xik"ş[VA.]{ݻ q) bh?%Z%g9,5Xrh{qڝ LWI8LߎKIIub apf3O߈aǍ8<.&I-DHdԳ#RϦg m̏ ^|jEA]RdmƝ"i)NlQ-q^zFaZOW w!5Қu*/{Ǎ=ݙ-"i/'.pIJtq$ىwŹH=[=#8a_bU)FJf C4 `u8<Nw {fZd?vzvVM* Q^44v)"%6V+ɴVr:VmA8XoͿjcUQe @@,39Oh)2"G C?P: $]gzp!k&TJb9&)b Tޱ;`{M˖66ҩg8ZOg2/T`0dru5^OٿCܛ&iv<+\Fا>ǬraNAz ;fYֆnדy+~mψ8e>4 Gt$n>btO1P!;A.ޔb푥X6J3kӯ۲ǜ:9tyF $.s&RP?"<1n2)%$de^h"q{;Ł8 sw~9`X!g>嫁nؽAݺݡΫ:CKQsR~ڶOhE-z8|0緿)zBڊY/ibAEny:=bj 5uԸ{)6w[6"ȖVM[mf_ ewkVSuWb>~l Ũ!C2GxFHθDŽB3E=}HLC-Jy>t^&Tz.XM3_-iZY.hlHF]HZ/ݘ@laD0q8CIФG|9M@c}/-H@Ku6 !r8ĎK„zGb}#\΍+fzՐH})̈!#.tvld_ ׼؉1xc.|-ЪF0~>u`5*qmYZc7Jώ5zʃ w5,Y"|N/޲쪻5[j3Č,J0m12{FhVMɳ#&&ͅ[O?G難j^6p_|yӑ8ͼ=#ڬ3'^i/+ഴ5pڌCS5t:83\u2hl2"#ƾff0AaLY RBlL+$ یT|j/?En}]=zn:,ϯf wv۲{CS/?=7uA+Zly% B(Q&ĖC*@'E'[,Y9Hni*o~$&gS(|`ɶ"͖w|o!i CaW+RhYs{fuס> =t||+g|Mr\1,F~NF9dz/ݝmP6a0u{${dbHc*eZz#&Zvoy<_m[9 * zO7nnBG=Ѧw 2t3ɤjzD6uĒJiK)+c{ɑy!_c Q+ bfȑ?.\* r"nXؽ-`0$_fOUۆo#_v Wiшo_+8q?U]4WLp~0] *N8)ī 3Bڂ*"*,h9b V#搲 ?j~ >hgGZE}pڜjk<{/G7zOfd+>\K|PbYpʀQP_4jrYd[>:u1F%PPI\mLY:"ȱWJ^Z!&na"7 f%ؤ1bFٗ ]5Mu” Jhi@>gf{)(e!V{79$%D¾`.6d? 9po֐n9yq\گDhl)}"0/IeǓʎ'Mvl:ly TDrx!KQd+tx49 J`Ha`bkGz9}_dvfTy}~4Aר1"O׳›f'I`O{R=ؓꌤNRd);5ԙlLdZ0hXՕށo}_ԋ\hGrfGl e?ʡcw`0ΝW0ku_/&eڲH(vG3)QY"i! "Y{bF{g.F(ТVA,Ĩ(t I%iP'Y~A2L=1z1뫼aXzzst{"A+-rvDh> O3SIqEk-,QiPüɘFy~ɢ2I-,?\`v.̴kPUv:?2#ޏ{E Y5Xcoo>gg}O~:=gϟ}]n1E,ggl>OdO;g毧r..:MsD@pl'4d@ks#C10$Xi [XgfُBPa96jvvg&ZiLVކ^v\C(׸8 YFU*΅) g[ P&:dlL [MP&='k V <&0|U7\z@ȗriS;jۥ!I̼L(1- 9bBvtraGMifyRt=]MJINQ;m؛ћ 4 klZ#[ %ü(AoQTсf}}43EDFVå mGSE@ Mct@H_)HA KhyB,9VTBacBJۊdjTL 'F=i_3_>:e)E)B*Y+E'zL9oJ`'6_NK͛BCԧ:꛸qq9Ng7翽}/rDW)Y}NH+siB4_\ME{4`7̦C;^EZ^(=@eA:,Le*un)S=tA ]]@QYS-wHA+xTAK>}k-WF4nMTLA^$Y']٘`2@9U %{;bFf1UIAy Ń H*0B0~ Jɻ#f":Lx)ARޕ(F(;b'fV!fC!.ւ޺TryNd=1sY QG7 ]Y U7R'VBI!GA3Fi rU4$ Gm\&PYGIɪ%)\`IVd6Dr;A4 vuB3྘5_0e9lS&tUP18EyV,B xn E9-졌c,iݬTP1E ')TCx6˂ lQ?̖CgnnjeB zoä))ɒɖ[8ȉrJ*RZn 8ނ]1%򼉝w3I:Rzgu8TSS-0F:|VM٫yRZ݂Çwٟݼ~µnLx>]vEYY'֟B3 RNF+ #nX_OFvxw}y!_~̷^csB,蔏o%"K <09]^A. p,gsm}I#fIY/wF-J &vew@jHo??iiͧz׋ud?}O^'2)Ҹӛ|Вh&kD%. ƸwUլYbGO`#KxSκ4hC-BeaqV,go'Ux/K r(}:Ǝv+OP8bhC1<\67+r ԯ Zqw՝hRKt[^G9$s;-$RtBF/Wt} oEy:iGu R/P4w eO_}{f{6dp/}ڋ ۝.dzӣX{Gߟ}x{z1喅hgDRHC4IyQa)KvUTm%Ýko8豪 mEGioGvҾ1({33I^Fǥ,j VA'U!kկ&%iDnHkT`I/5~;}fߛ}tzQz8A.=A)&JgrD%fIV$$Kdkt$d.(sv΄A^VeBo4eBˈZ$Je;3B#q^"6Y=Gx{J&wNY q+w_`lR<,ks u@۽qxeRbaxzS!E)k8aot̋AK4߼𓧈u2!;kTPKcە $׋CC+ߓM~2ݒz+Tǟk?z^FmhbZzZcO8/!cmT"$j#:=$(ТAB%~4gJʟ6 i1,FG3J8V1sBs=g4Z8AC?ps :F4#E=Uґ޺-"=kϺ.g9){9^ujZi^^FhG\rw!/l-*m~"C:CeIJ (4qo2ǥ=zK1/FVa(Sc(id W \<(;K 2H (vbOR@Ѻ# hvS:G`[A:"XK3 yQ؄#IrBAtBGf@m-c+ZQsce:fSv3)ammZhF/k' E;ØhI"f͠@ȥ|'ġ=-yk_E_3}oYm_\5jCTD[\JUnM+j]hFjbQ˟]?#-mdE% / GVjN;_;d!~7HdHPyzMT E`xو[,l 7Q22Z3*.볒m͖T'0nO5Jٮ/VЫB.lxû_vO6X̖S B~s_g2kTS T"0ܬgWEY"`M"qB?!{qx8s|Lǣ[NÃpU:$8wG sӣO*? }gc/3YԁfSkS _!%;p BVQ9ZɯLЀ2g֨U2C$).h<')]* l 96f׮KJ"0zQI`Dq]@,#ԨOi/uFj;J#tP,둇1dÇm|J,absCi 0LV5o_Ә60w?,zB? c|w0Qa> VUX.6D[Eۭs?ݚ҇I߾B7;Sum(7dgWw ^K)\z#`v\km; oJpȡelhRCa{gpu_V53o^.l()|jGӭKVb%h ;v>4b_*T7FuÒøw?tp9%{+`Ӓp0hGX&=,ZgD էh ZcNb tEfZPfV?E!T.Uͳ5R:c*]sKR/K$LQSje*h=$,~#7=Fڼ0Wy䏛-k%]E62bEl.+J5,t^YE}&€Iڟͤ:}G7bEj07=e5󗃪T%=K9xr4WRRY-ԒTd(6Eu 1b4"YaȺ;HH1BmF*Ѐ΢YnAJ7[ QBfF aw܈`1BCI҈㾥RbǼzf} x2F+B)lBYHn$Glk oTTM%EkkL2#"e2 1íW"r )x `)T)VF^v1H1QȬ]亍^ގ'_Mh5SZd93L7t9C膙C;d\sS[\f-nb-@櫷L) ֖`inQ Zs$EfcU.L3]^ b?_TBZgVh nE$ b$#fCrD.L׀k3h0isןgpuiWpH~u?I1Ɛ~}`.σ]rb|u7DdXTsfskXsEAV΋>Wn~4p{28 bWԁN3ΐ f~%1KnXBGYP&eΪq ȥ + rSRH0Z &vz3X 'Kː(F(˜/d=u*HQFJQDt+1JЃƔR;)Uryfb,H?Ryq( -&c#h/ѐb0#E#ၽ<+ fDJαp̑( BBat|ΘdI7sO_#6VH^+GQ.k5*,s[AȱriDIbA/G߀QF k)jg\KZY0Θa ޺{TeA/%LQsd_Ah&_oNf,0S[$04KB. mgZyx}Mu߱#7%0s{4&\զǑ<9/8G+%;-=*qaJh.Blj6CospT`AKT۳Mmʹ֨e#"ۃ&Vаu)!ȭ&%E: J P,gE;K*,7Jo|u'GȴU1Ă]ziW-$3X^0:U R&GR@cHgI֢llyO]gy hi^<iyMm8^O؇pZ<h_F et2&pB3nry)\Ճx)Y/m,mO_G;ws-ZqLYQIHKZᴭst(Y>HfxK/n@ڒg8H?gڬێږ! 眜72" -k]U.|v4\Ђ4' Oh^RFx6W;9>N>2Tl-uڥA22bkU,PT3`5/fS9O5ĢG kQ9ڵɨnoa19Yhw'=K+ۙwYb J._~n獶Hx:Ogftr_FooǮ=~LW>-]k ϲQs?ݻד%>K%Bd\6k568μ=ZAn5yq.yCmg>}L%rL 9g5}),cYm; }N3NmT2aL bwZ$n%lO5*YCwjDurTkjKY'lOJcv #韛-Smeb6J^ 7ly0k<(PCb/%9!*t3HiR㵣eV3>&okG8^HcJy`$d^nk|[ bL'u&mw ){nxޭ y&dSU\vܖ,7,eںk?-QbDW (B{ w׀^ ۀ 0}9qxeJZNrM;? ?7ӂ~`,L.no5vG,k9R==ԲR5ԣRRXZ])JRJ-?uJi%*!ƟZk2WJ޸.Θ=7YYJup~}~rH YHTw}}ReRAjݽL9͵n|Y?X샃jh#9=:ǁlLk`m^z=J,r\% 7 ز՛E=G h>z/L?`wAۘHjKsDO}(jH7Jz[q/c_CT(RIz?vP:\dFnUE*s:_6P"(iz, ʛ}tT 7И7IZ+'k27Hrӎ掞MV;䅒7YO1@ږ1gw y|zr.c!8S7qq|is45wD÷t8ZΗ_HIC'D(u#͞Y ~ŷ׽VBXgn#rlS1C+,Љm z. ^ yÁ@Y8$B0,EW]^[y}^BBޫNW0|b͒Cu Қu9S %ly=S”!Q02bq[J4 UtaosYnnhۨnUp66Ͼ'I9" ,M[iAW ?u씫}dUT%[qwZqEu`KinHX+oJ6<3Q~Gx ;~U{M/ސEp2e ʃ=Hx[=3d3'Ok0:ZwLaj ԵQƨ2`v[df4Up=L2烜?.^%L< +¤fb?3.`|~2_5Q1X@xYց\Zլ:}uM־X}u峍kN䗞xԏGִ3zKO@|N>JE8yD@f8*6iI ɛs!CƘ@w5Rhj QqPGk/ב?Z_bZb  I B3F[0zΞǪg|4 {}&6^ F8KܬHF+Qk n6&;-cF2DLs th(JYi?KY%6M᠀N6epQ>`k&bSd.TT[xoVYN+ e&2 `dE83w¶ݣInb8 D.x|x7N<ܲG0*Q3'Un L*:fj7?=}>A# /:hPˤOkq[ndlwJ.~x`$?^2?'rPRoFFH]:B *j04 ҉h-2,=PRs޼&719CJ2k>xJ{nW9k hL$Kƿ7;ǁQlP+!{ĿSS=Ca!T́)wi$ &;NdYeΒA'"晴ީAٓ]2_fScY(Rqe = +Ѓ7(n1ihf qͿ~MES_.ꢽ(C]i|Fdg a&>R@LXǶB2SA.4".Ot"zf0up1ZY=yAJaSZ]sgw6oYFr=D! dFA2'J͋m/e\RM*q CWvnKFGIyТmdD'odsB㸬ҘJ50˸4+n In+L9щ U30c^zbkxahFqb|tB5$Tu'm H\>yQSPˡNvBIpҽix' >;̎ z%gsR !#+I8q:)yRgI Βՠ]]1 N}MF`Q1*2R&(/k8=`:0tœ{KF.`o`˰9ɂƢr\kg.hA##  A|9BNL;ɂIm_ԥ2C٧(Sf YxX,c'@)Jd:bKHnM((%RCLk-im-U|gο"jH/"j$H%քlVIZNg h@+7l,Q8` D-_N$`O"j@ 7MǾ|lgX{b^x{4֎:tԸ$(Cl 7ovM{Ç74f!"8N#[n/ [cOOn41PkN#|c>m_U7k|%Ѭs=ȴבPG#_l:xpV[ƽΆ ڟkd]^/9DY]r-3UfPD[]^_*U_nwZH X>dQ܋p%@Šh"wY_/E7XDVǵ _  fgqt菛`z.[.#| mo3 pzEI*7)CLZ[K*J.]|K:23p.rGed3depSd,Mt[w<"mBDKa2󗦢#Ϝw +-z3sƻŒ ܻ 1;/bYκg4h-}"Lu3D~P5EdTΚ3m]XɁ>F͖ܶ< iT33O<#I&Wّ' ( ޤ3 :8{&`cP%F]~dVa`uʹ ![vFl+]F{Vw KB-ѠtFu]Z4>&gѻLx)"/0+gi) )J7q)?$jPZGu+% A0*sd"_g%d,ݰfB"R2ӡ_YSv;ޖG)ba7䅒ڴ,2_+!XIcJb&E#JSo7^cCV,ĚCiI`er_195d\1ݺ)GdCy?U*?k!Mn}ù` R щrݒgdIKy'Xį ~FD;n:y1v%ȑSf[0e yN3.{V˳ex<( [zZ׻ru}/voGlݾTD"'IIΒ%%9KJrV7yah<)X^āup I"IؐL M{XBT()# ߉uNyF2Pj%X C8hj:T5PqR9o;NҙȧF@\qoPXd+'4v8[þ6)6}yM,d` _HW T@By*zVEe |ŧO23Wm r:j\oB bhh:F|VPMN;=돞spzV8f hsL.zR(qx#Cw.vHn 7-y@X|F1*i|ӭοcEBTQd^9#lTA\Ձ5%xG*6t^S[ ӷᢖ5ڇ`J&T %=E4B#kDu`'{,'fF6*,w`u<0 &V UV!VF#Z8^Xd63@86iҸXͩZX*4 EǪ!TdUxbqpxZ0y(hի T _LLvG3iQǣ<;돍\jyRvV]SȜ|quvS Ym20z֜:HS;^g4&T1~P"qJ =p돚! Vl$9\r1jwW%|8Bq77D{Jjws}_ 獅`]e/nKD =wt[jW < e>ˉ B4hOanyo;CmjD!G+'LSs>{f&OKxgY ܫX1)\וi#y ,ެ]9a%U0U>ETG?-58\#y&Z5e?|@n2Fj,zR tlpƏk`ݜmӅZs&2[o57T)`dƎ9:m f͟!@{s!0j~`dtPje݋d`jLFBX=v#!ՐK(SA) ;>=POI"dOݡK\U t>]'Ǻ9ݰ&;%:uT5chxXw``ZiK"fUDW4>2%Bu`'̀{=q+?38@Ж*;DhkWLCw7k<1Wx0iT޳m,W( 3C8=iI_=AO ,wj=Ivwl)JR%ggmBѫ/sm@e*p*@3UB$>ߞ?^|MU*pBcA`5Tin~)Քry}/["`ćғQ7*?w'aWdZ 'Hq? /f&G?ŴQou)<ѭStb9yy2c0ѷ$cV#3k= z NQPcDD4"\6&=3Pfd8[qFw2EJ~s,r>z'*$>[ I"f=AV ¬{)}$Hl8=Y؀h0-Namq\޾*JLa'θ<SR׹11YH|BBF &c{%f&HxEhE%SXt֢1H^Ep\f p\n(um5и|i"3@Az(` X IsfK/Dԍm8q<ǤDOꪧR%1KĜw(8)Qf8"'XJWQ#_(tUOWQ":., tR*I'\7P qyFj9C3]6Ρ!8/|s-5#{ .,95ZVSc9$pk-9Els,p!Diiy.Q )]GӞ>4U^ۙc aɅr˼"|niU|"WI?TL'vr16ПƟzƌPE p2j"D &4c0VSl(GZ ɷ68qzlmߖ'Z3YĒ4$R)#c +$v9}GI=fH+&6Db|&eZˌ(\_| Kr_jNR /ILԔe,I}Ƹf+83opOaǮ (梦\&.z؁2.>\<\EKB_ثp)RȹrD8䄯x:u37d`(pKNR?cWnntt1-0C4vFU*1 !o`oj$ r9KUNΏƓJphƺ"|ǰ$zzGCK0HsOh"/0,DNNN*@~dRHgՉ 56uh;.Swf IpB˄IU c6|P@j6\_^*e&@U]RMk`%xD#f56Ze 4GR1_Mi($vR# 85I$RQ[͇&-07Ȅi| &%ije@0'L(-iȝqÄԨ -3B[Uȕuņ܍Bbډɖi)NY hf@,\8(^IY$/n]XZW&.gyӗ=guSzZk(`.[wiui2OT/_`1 3`hK/ؼzUjY/3:;|XbM8z%.H $DG[<-."hU? (O r6!w #f+$BCWnu8rrI 9.l^YTњ&5 $.] %8.lIjnx]#y]Oc6&WfT$D`*=2h얙֫+6/E^E&`-Pk]DEbwgע+J }P1 4j5u$U{66G$C Iy"55BϴYFgt0"\4nt)Ro8k$5ϥuƝ/(t[1:d[|:r\*^Zi]+:r\"B1spĵҫ5O\[jeƺi?e"PEmn͵QMׂ&0l6䊚.G!RW6Z4!{ս_A:R0 t#[P*PR\2kWt%+ERb#bp<u?/("(v"*Dލ,v1CB˚#?/;pP, f# d/SR [as3j7 Z9U н_Mҳ@?$:u#O;,Y'%wrk% 8RH}⧶B L[?@?w^;Q{{o8P'\;8j'_7/v;;<|M~ m0\?v}ggΛowgϐE o_f~޽~^:?|/~B};ʋtooɕ==k8B~i]/֎ݾ2 %Ƥnnoe3?2=~lP&(=:̫ޘ_ ?My#PzU|I`0ݠwٞv~<7gw$sY?5{hxjhMٯN=#ϑ.`f{ʨ]r"'f^0)~}Z9}~m81vx.o9᫽Prޫ٬_<af7nc}qx6ALw  ={67a ud:v׾n#xe˶p:WM7|~`Up$oIox4S{g@aww'y) k{+sjn]KPOި{בwSZ}];ͣ@7tt.o_'~J_zO쎭M7ׯi~;l#x=qq&Oᾬ;[ٗ/G׾lL`θ{ιf@:\?TAθ-N_hrv:% o(\'gNt"A ^C29ȵ]n¿;h[lUP5Po3溠 ԣTKqOAs`"e&Sͽp)w ֥i1`#]%]й肍.肍.<ݡIp&whRtnH \ \(4RǗ\pvk|==o3U0v.w+Gwٻ8$W|`-:zx4 cm˰wI!O2lɲ7ɦYnWUg_|1k|ut#.zF.Z_~Z}-Py/ZrZU9^SDE"WHƦ(,8B/5AWFh=BZz#u戍zINv4Y vB8}p} fe?Vaa9A1°NTӨG$ xG$8" HpDCe1L?an! `4ӄz@{ $1 ͞!~tLk[ 0%M''7nf<Ƿ],6hcV3{-pF/\l뜋~]EzIJx dSa"%hLHBJtK7!P9S}_lAߗo;P} ]q?5Cbɶ"BviRTB"gl; VG"a#1O8sOnyJbL &6^xL|F_)Ǚs ́!=x,:F}@?d- Wg"XR&՞kb 2x Ύxt~O x9o~#ۆgSϦM%?Es?OǸ _ؔiT Hڜi#R`X \w Q\Hqe〿tXu *Ab+AXvr6/^8_bKziS6'ᔷ\Zs{NSXJ~b\viK[Rk1)䒵I]Dbٲt\Q5D$FzC-lDY,2F^jR ) RWBt_J{ U>O)3իWGpΎvBACxT=f>:#::#ёm<>HI~b`زtca+l%׀ 1YV%zڳtAْe,?C)8NԻeq P[#t5 ~)fö$T m5@ hI\h`N:\q\/(Sh]ՈL)Jv#`-qo3O|Yڰ4SoRjZ,.1hOhIΣcmTQB,-K_,daIJ%^B[peYވ@&LQ+dJcpJl)=W GC\d]/r3tL,_Eo%Y^4bx'WP0J0  |鲶 h0VALA1HnjrJM 3\izb =܃FhKC::.A1POţ&PqW//Zm$VmAK 8 kL2g؎pNx"eRbi{Ml0X;:ςZnɣI5r`K憣 2ۯnμ*8QX˱.:`Co`┨C0 @H|WE+5jqkXeٮazyZuMpq''y齎=ivAL볖Sx;ZxI1}@d۽nud',ONɨ.@-Rio%] 9ZpFK1ǒ6ڤx!8xTB# G+9` Jw}%Wp=YgdK⣿g7غghV|Ƈ,!yd YјTCKLSN^J\.,D,l1n$+7&UnN;ֲljJ#氛LjANDg}'X\`jJtBwGbMMo'rW`uuf6ѿYZO̙fU}\/)77Bfk]Qm-yk0oKm󮥶ԇG4u_/}B[r]k3b$û*>ѝN4ʌ#c)FYG'~B?Sib3G' +)*Y׫ncN߾.@m^Vf{Ե RBw~; `tQk[.袺p :]dDžxhO2o*s͋1ՕxG΁]>ְùb9k}~7ؙ %Z aGnkf$ùé'-9v&XVVoG5^3zm]Z1ö"UMV.w֥ =n kۅ^N q9*؍C{\ 9(-9Nk^<D8Ju 'Ƶ Am4D"G>QQIad_)='B'QrC-4;u6Һ廑>>'p9ʙgIaY͎hG57Ro9<ıA g$"fX19]|̷]Ga Fdڕ':FR}, 0}:Jc[ɴ-?ue }P]DVgD{oWub6gGy ۝^!ڨI8i̼zyoZA/#'ɷ^S5?jw'y6!to~m\S^]ΜpJڈ/7>1+(&fQ8V9kߎ-A̟L: qSAmOgP3 qfPo5b6)P\ ?RWΝ'F2?=je*UK/┱οX+sǁ 93.;fnQW.cy! m'6/&WlFI4|mG' @cɁtR M{cFm8PY5l캸ӆ`ᔸ6\h =x3j f+"(ODL2>hG^*zR;57_fKop1}l-ul^yVI_g T-˳_g3B˟^◟v-Pe'4˳Ŧ׿0GfozKv?:".|V܊T+2@aж-!~8ice oN D MbF8Kbo:wlfևe ꫙NFIs%72PmK>  :@ͅ' jn%Ӳ0B(iع}Xw)źMwoO>!nuτIxYRXϺ|2izlZ"slu'gy>yl;.7]σ(6۝cVDzCOpw~M[ 4gͻ*"ۘFSfЬ 54^D'rRc)b)!$$\+,115J&_%唂ΉDnuJ k&Z,J؀ZXϏ.QM=WϏv/R8C£3ju1\|)L9EI@wVڠAK FkGMwQo$0/濺i^;饯q.0 wiro~{2~ ~-ձ`'6nBGֆ ]͞-+G;{hTZH\O.7Mƛ$&M=&mé]OUK\v Xq4QuUWU J:[cqX>n7LF ޜVBp`X7sQ[ c@qƪ$`P!& E mXBUdT>%CВPfE\d,ROXN.[J`Q0f&GN{Lg-A\$=5!EʅT*WA3G^'EI; 99 RR c N"ғbOOsExEmeBۯ-̇P:Ds I.$a#^jܱcJ99mnXhj>mސ*$;'h7<5I.cԊvغ4Xp@ClS;xn{!naFa /g)qKkrrɡF_\6糋I7_ZOoOʴYۑVyur}hra)[^8o:z6I't2'˿1=qzJ^nv&\.xkp8iϏX4Dvj>cx_Y||޵4q$e/t=22 vzZ\Q$MBcbf$ e 6:T_{7*-%W/zxߖ׋3.#b=M:;ٔ|Ln۩#^6k`ٔ%c5zǦ(RjB,]O aF )dvrnr30N}sM& mgy9NASKTnURLd'_[&"'FyCAox3ʑV^J[PW ڢ ']Ӷ'ggA??d\zunn 6c3E^;Tf%/rKpڌ|6.x]_ nZ^B-[כZ/n;$U' >9z[-֥M@t!U^+QXlYJ|r3*Or3yE//k.F̪ܺkYrX~M%ݿʓ|41o Sl#jК|̪#.) AY<>fy/L#e.֎ۛ-7]D':ӑK 8NI_*Z))8!AdBKEXR˹Xd.3~ {c=*6ϴ9G} ^Dl;Z,9p9^$mb*!ݽeWu]`QJm]NZ-5߹g]Gj<$eOv%J'b"̱>۞&|5ЅLPca] 2lpfP2).Ea7Xcɷ}RM̪0ejƄ=(c#|=&q@LjLy|pۻ5iTwr{^:-f5 ,FP?$0\,ƇZ)fH6-<bޅg7$]0sHȨ#(% !IbP-4Z5b89TBQR`-D̰=25'%Kc}TDX0{R JFS[l[ڢ@ \ ,[y: ʕ־&*rf"i*z8vsuM|k\co˜7*2[׉$]%* M>$&U3T=\&êb: UR:Q9 Ae")Zs$;qd8 [\#bDx6Q4l#S%̡aY&l x*RDaf)$~~$4կ.sv rEMDRHeЌ{Wi8Zey2"WR!ObrI"Q\u_3Qޞ`K$f8{hŦMiإ7Ϫ(tK3$ leiwIIX: Yg\c-C ki;xiƓo{)h9n"dm᧖LZaP>Ihj !篍)݃I@Fy &݋ɖEH<Zp~Po@ Lp1Z;6K9]myqa3"@"C)! ^)mVNj`D>d[ؙC$n4H|,~lVrctCLkx&&o 1iLVXɸUX&QF,;ªӜ"g2mr$T 4Ͳ[Lc<!piND\M"L0R)aNsMiZC [4Jms/犐D2t@灛avs<rjtX&^wSL"nl2bUod6ߩi!n6Xjn"ڠU->oM le7 S| /7sn~n55gsssssssss_ڥ(e"E}aCo,]TQ::uvVSmgzc dBߝ}oT@<`3G ]Knn_ CWttIIPɇJ70a"D1 Zu>lN瞳:kI qJY Y Y O Rhù-~B1];6Ձ.ԵI =QC 揯ƍ:dG[ ߏ{ġTm*XTdŨkΊ>2#fU䬵Z n!N♳:+I9+:+:+)I}t+vޤ^3Cvb ^>}T7?Cxc;?Y2x_t${iҍ/Ufp>(c˭ɕBuų-^;DFʸYV͢ pY%$ !ݟ%Y%Y%$6G1w&C DʑŮjvVպxJ߅ +H3Bw/N¾Oj6ݖ) }1=5n_s۹כ?WI\JDJۚ Fe~@,SKH1iIy27ě壁@:Ic۴>S֤BELS}MvGCκ;P}XV C+65sѲld%hV"k>jcUkN}\v5oԡ Vb_Ok.\|.Zu KM,j&"]r# ا:~IZ'-Boo>Sz̴1} ;]vˮS=I0"J0W5LDH$& ڢIL4U+mHN12s,Y%U&!>=ƹ{Y(Cw>Ub %$qK왴D20X t )4 4D㶴 6bGDBH2}7o߾ s21&i"@nzC] ɺXc:HCdGާ$*IX /e/ E<ɨiD{! wXIrӅZDt?h1%+dN BmVA 2&z`@yk%U6qI&.Mrwx%ʦ)6Tާj:zA=-#`1d&tdt&Qj۩9]$(E&Ũldvj1Ƞ;{xJ4k0sJ)Ue@)sbZKk)1gd~ݖ }⨰wus͟g~p Tj=`ڏ=tIx酪6to?jF)nv\:}6Wp4Bv@P8}sMɊȚi-V"}kȷ<(ƺN;0T;+|q(A@ 9` '\SZI VcPsݧSH<ǣXzsr )xHeN;/sCl*qݵ6܃Eڵ}4 lsB}`-~/a?WQk88iM7IV⥨(7czi26᫛Eͻވy*wˡ~) .MkB+>N\ܾ_ܽ_zNj7o8{ˋ|/O,3oJZG"jke/v6jI 'ݫU<>|훼zM|et8R0w5L5hრE$GNz{Ǜ<^ed^14ShTZz47++~975"|2C4=m]m6ic;O!N_n_iI(3{M5T)Tst*q"T & ͬm1 h}~EzߥFxt?$>i( O&꣬hGAoTf$*Prf8)yȴ5s mxԍUIܰ:֦oe}[ϙrFb)e&.6dcAikQTؖ @{6n$ _ux6\w+UGm.ɇsJֲk ii8$% AVJ$~яyugu7E%Sm_+#ƪIj˗iV6~@=wڑU+uk5d?>(q/O-0$X} !RT3Hϙ$VVi@ 2R0Jjř֩(δ t39D2Jqn˸hx2€(rP &<i[V(rG49w.vbQ}e`5f I.S֭{$UI۱l'+0qԋ[[_MoJ}f4P c tv_dZ ^B;mөdu`g}vnZ#;@eh23?D06E%A{^~|ote3=pLjm@`GArAPwiɰRh Z:e1ݰ %Sr狩DVN-:FəI 8v~N("%Ӝ\(e㎇{gHҚXXXi+JGMYz'K.'մU)pPi;(m?/n;oϮrY,]t^K}v{v^܍q \<1k!} Ivj:@ P55ϐIˎE3<\фyBǹ4ZYjpRn-FW ʋ,ݥ/>rK>m*f:n=M.2SC(p32}xW r&m^I[9E *Ks>LSȮ|6Zhlʧ*<[哽^>ϗ_jc3g(i9W_UxuW'5Zvh"9*wz`(YwB9y޵UƔtmax-_Ni|o  *,t5uLn5|{y<ǡc&?^\:Nj~ pK\ @,X4Cdh@^4!yw/w/RjZ>;G :u%"AOy>[:_.Gg7z8o!|6O:^xk[,g8R?nԚ6L$U 1t@[ԋ?okb;gkj]DӕEpidz];'|=l&Ho{=Od ǟ d,.S>G%' =wA-8jrtJ-*hה`HJ[{%#unՂ-"`}L%:V,~B 鈋ɼHQs?y==jNVE `fa簑G"7HIK{njSҺF4#WJۡAc(rEρzͷC>$3}j`!"-TQyj-j s1Ua׃5A$d ҭWeG[&/竿U|X_ Hз8H֍\]mp,-aYh*ܶpX"eEW܆i`_srѫrݭ'YxJ90bs`v W09wU0p!x%跓8K5Ǻ>ښo7ʌ76_$eM*cQGG,9g=Kjn =., dJYJ&TI(^ s;.I吙k槊Ȭ%xug!lMFK^f=7opGfջ3d}5;7& e?rR(=h_. 4Zed*2 J(c(T YC{kǝiG.gǮit6'M`v/uFJJHŴl%HJ]3dEyFN٢,tBrg=xvzpL}v3 iD.eKߢwߢynOg,+o8}nB*.?W/=u׫\.<Ŕ[Ef8iUhE Y1@TۿsW{n`1Uں7ao\t?L^NKW?<,$`VU&}K%0"$:4qtkn}kM#{Y{ٮ5 V5oÏ4:FRP?(H#O\Ou@a\Z9{́]q9*B?0 0 TC,-"7!z!B7^" ?< 0B~&xD(%/B՞P i@V/&2Ur -űu[av`L.wN2$~i麯u)8[B3U @cw4gzkJ74سTFIM {TݷIno}" Sʒ'bw O- POSmd&%TEuA?g7G92#cb7XyqSb1z)jEQmF{ĨҦVIpZɣ&=nlۉחXt@*9}ʛ^Hム 9-{퐈f9H]l™aBjl/ʁ@K]=Jw"e-II^29 kU3Nĩ^O[?u@+H= fP&x3c(+z/׊Tal(N+Kb&xx5 `.X#=j!öeb@Bşjwb:VϏܤt;]_j rh$gLUU&ɤ%0$ oVL*{%NuTjz q*BbtM*k=|b<.Bτ#iB.2?E{<NԮX4D8V=}yԴcG5x3k#I}Ԭ'੉BPOR/&wĭ%k'零VFue!ʀ~cy9Ղ\Z8gS%DM;5'Awi:fa3^ǘ{`/Ć|h?szZ9=oЁIg8 DrTkOˣ K2GYptZB0UetS_ P]fFqTeHS*)\)K$3 gmu di`9*J ̚TV$'hd;(&N42 r'"D&hMoaZ0Sޙ|'onD"I^ffgK}(v힉>f^)qe"%oP,RZ$H)$aA=5#aWP`w# _mPL%/PCtJ:aejIaA!)h")7Zi>;)*"f ~~5L\`DZ[ˁY S;Є._*R c 1-z<[SBЭXIV)2:,CVrSz5Y8` qWÏ&9 !5I =Gy:Ȉ5Qf{0"8aDh1g/:b']+:`#UQ]|x¦А= ߵD[ 2kQ [aDB3g2Ւ/v$x@T`yˆh[1II%jÈt7CiXݐliOD:rLKi;xoKuU~{Qk9B?iǿ>xf%(ёV_5TPn `o}6#8|)reeZ<,_`>f:Gz8{9 ;}臦]M`}#:ɐ ֊p=nۓyMoOJ6A9p$F[ʽ@.>: (hjKٜO!tûS-~ W4?}.$. ~gnlB:&7ǻ ǡ?ՎX| {Үy&=>X2Q Lx!g59q-G@h ٞ\X _D09`gS]/,%3R>8VFDxJiա/(I[ɄV!I3DȂ,|zq:,SY~<,Oi M7~ Az3m宛Y-Wr5;WF4$`DjU߼:FS z/PGcRߖjBp*TQ[2A %h"KϱpJ ,QN}N2;"HJo5Ex4 ћGb!]6EV; XZw>BaD[JU4*Yf:y? 6 _zRz8~-jקnv=)/g,}4iʒ* yk>Z;G,R7y o1))Ԟ 93͠C&qn2@nFcF}?^jӬؐ'T%A~M#()ʥ0'~vw^oUPnZJJ, mև6S.zHkh /gM7R&RIv#Ály 1jX2mK6r6Q&H\ބ}MNܞk,y%H3獳Dx3\Rg1ϑdVz#Qi TD;%BeCLp륒q@쏩гh=HiypD^)=t&` a=@Y43Փ=͍0Z{^&Í/3օv;%3n3NY?a=cs^sGe[E?3@]*}+(֥2 rGt͹r4R77{ Bɶ@h; dW ю5ey3^\[oāK>Y9MJ5S2w{U":Voj хzη.2w\3=Gޔʐ>~37.K&BNsWc9 Ji"8א[*5#',&H{I8G9B[ERH)ndcpޕ> o߉zk~,f{~ 8'.ݝv;C<ٕ~+P-)-bٹ} i7 J[s/>u;{ ,,~w4Wu:1,ԣ[H2/j8 \ S݂A$ S=CL?O:#e|^ `&q(hR;q[AX( dJE  NEt|W~;x?⇏XAǙ56;໙=OA IԨb0+‚]Ny=ꠑ^#&QasrAP7;5 ٖ5\JYJx 5Dp=*V^La!%[z>T" i6G3iNf' b&h:TC"hnAo3hGx1Xzl49;p`ief4 Wnqta0XJ`Mp9˝`s.1c4sL9I ?A(dpbj!QYlဨE8棌'/rÐlY]V0 ߖ1J4f ɿmŅQnbsHd8fjdB҅Ɯq+xAv:ǟGC⌄s'@$ @/ϠH4pWg'IVFKsHVmDFKo%uP,.]"N4e+{d&3Pm9?{ۛ1g)޸FqNVyUcdU^J^%I a8 ΀<#Lic20' ɶbTOc!O 6}F1A!էzP;J><yvv NJ`1;crYsصD߷PHۛAKME\|G'}:]@B:prV 5-^u8EI-|r74ܐ׊r v=0 ʹV(IمbR?|)IޅiV\NCːn_y gkH0JDB /KL“[px-ǫL|vz'Œ$=79j:uzT1LrF욹pge-1ey u&)Q!Yymʦ[MTH Et4Uxh* s\lvذa}xhKIi4(~ihZMTaEpiz7|$BP*ʀkIiP6bEK$D-i}w2e{dIR%póC)?sjWK'wL;*'nRS趏r6>݈j5i<ȠП;YI| ѫ/?矿IŚt ^Z Br3tz*Ј< cYb"?LW-~B]Fz^I"n맑O'lı-'~F9xE@*wzxA(1AQzr!Ȏc?N >o?,$AiSu{*mK>Rﲸa$ݑkLĎ gRl},5$UD[ݒC8M!SZIĿŌ NWq*9(UK>@2Lq\pҖ/BY>?p GUIUGWnN7Q%)4QuI_B2je]\Φn[qn[qU6`~-S"C08 ͍ErHD2唖9jm&BZjHЮ[E"B}jwTB'yRʦ5+c3* x"̈Soӿ:v6,39-OO^}לTjfo_VȆPx*x]RjJ(ש؊c+>cQ"Xaʩ(-H )qJ('Rz % \krLVZnFuΟK-wPwH!*5$Y( e2.MJ+ϡ In?NTX+Z1/֊yV1/VT /ABVJA` p'i<)<p)({ L'$!.xKM;HElyyXHXWsL'aOTrWTr)VoFe9/.c{}}bI_)~7 qrX96=:?pm)λf,82\W͛?~QNxW3P ԪY@|cc߃we馝 SMԲ?$^VQ'C&:\XXH͋9oR|`WkјkUV:JVIݾ5t}5/hnσ@a4`, nn h˛`(o?~yy|>o|:xɸ͒tp_fu|_..߀/_Mwg翷N;"/tml>ik*7>4_ϏP9:͏ ;7vsTsH|3o-3?^G=+ˏ.|t?]0ǝjsv-nsᗵkTȵms!\A5+ۊ5k[(Z GmB|+ cゅvZ5g_nN;Wig4}S~5ȹTp3) +.o8Еqgus% ֫7S^K5qI5%xV=nQb0zQuM)M~Qc{ +r+9'/1-]fO>m9{:1^ܫ뜻uo>Cdx|~3,bmUps#c/_VP6zqIBsY;!:= !8~C \T%@s\XuP̰c\he-l06ʠiчn-s5vd=^R23jW,[XagcBkB}^'hgg1=q> /~()Ɨv"6,U{Vy70녊v{.~gEr8<ɤ܎{\W|q/蛌Wi6 æxJL .1NːLlmIGkǾE3F?c贎%/S-K_k2J˜A歹 nLV+cLZ>j[+ׄ[)žYq%qH`3SA0hВ Kv֒,!ߙ_ cE< *dI&VJƒdKFK$GjG:xqz6kY<-1FQGNY*Vi܌RSǠ{83]] QCXY+V~< G㩔,Q=ӕ`迮bZ0~Ψ}{-#$Au*h[O .bڊ[@HiױR!@-'ڜntޕR0FveW1  4%`<(]\K?Jz &3A?s,j+Lt!sKy󳊩X])(qS4" r*:K'AR :i54Zθ'7jL$UqC8e6;лѨ$J#*Xz.bĄ^i0p>p̔r </[a9^d -bJ.`EpJ"Kb)‰V Džp† SqelQg0UW{#d1_Vg@Bo՜ifeüYhd<xOqTjo͊nel)4%4&f޵5q+Vи_\凔:ǩ(y٬ `dnt I9n!%HC̅rXkݍXi"~vt[@FoP6Vy)n <kP `T=E-Үb{Jυ Jnyeh^S(sV4a))9J ҡՔ"0{HH`ީC3R"l/r4S`C.P 폐S` D]r o)D+ ޝ/ՒXMM.RK0 @ڃEq"m3'pK t0 5fF˭ $qB6vLa!H#aJMz$kX+RF3Y)o) PLdr!wYÌOVDR» !ɦM"Oр= £{iISQ/.+GH"m*U@ E1bca#& ;aD r9[mw6PmrSDkي.5rk]e,2,ϤX [anXJn>ˮ_z%%v~6wA'Y%1g)DPQI^ !`70jfֻ{rc拀/rOBp7'!pj ٳ:dbB' Ho_G:R{SiKUҧoW\]e]PԽ1\w$E儯kʢ^]5m2R (w{VJ(뮤>D#ǵp}A-=nm[JFHhPh#-] 0ؒku]q܎7@Wmxbj[\#iBp{oWW}ՅNj5HdfyYUrr"X|Vj<28o!Rx0{Ԓm?yByۻN-;}HTŸq"^< $ C$s7ш|ԋ4ɞ½UKgsw7NULLtXm_a7V+s-"̓Va ,r/,8{ʨ.^2ӽ U2 {}" ]-qL`e|AZ8PVz?/>^n# 9`^cF0e[g9MoaoG<#bœW |D _,UCH q'ɝ s/ * rԱqF@1&F'+r?D6]URBq?1J1D}[R%}ĸsibFP7D%g=v{ /?Eޙy K~]u/MOyʱ<[ԫ+Dص^xsz~yېjhyl*\4*tUs~,ܾ(s;VǢ̽enրy;rS: Vup%4w=2m)BvTp2I]xK$ t5#h@xPNmD/ŮkyDQ¼]i"ZtPygy6tvaP}ňV<~̇*!qoCT YVzR_b/QV_|V Y20͛7˯Y>YV/>`p1Һ7p76j,flO?_//K?/.+LQ8_ [2fϔ5:ӄ0 Z8l 0/)m0N+{CV~uLkD*Gd^"#LQ|DyД(5XR=+r~&[!CNuf>pJ*VƠ@ :hJ&u&<}z]'@D=\׭H GpÔ<![X$Q$ \1\ gtݍ0*ef-HOVL#˼H1;$z=EO?:8}rIl}׳,>=}7a%-i5-~AwJXCZTNj٪<28[^ & _ˣ=l܎`bI+c1nN'nœ9htu'xu[ Rc=^`^ai{N0J&JӇgn]̖_a ,o.4e 6u( 4[tUy|| s of9n,W39B 2JYapO nB`Uҳ:gEZ)'-)EPĕdV3z!'ď +᝭HڿX1ƯOӢB 伯^kYjɢAqfz|+a1%`PKPa)tJB{cxn0Yisi4.PE',eG8 R6=4WRbC}4%lGuhk̮flL ٥|(.?zX%lcjı٨B&| ZB.'TNtK4!~T J4BLzþrA9xfX3ִָRzJ:/4Tg4X%/BCZRTE羸nB:y| 텹;y/xzZL چlwO^ջГ:Nx&ڕ,GP8ruLWx-W&Q9C< |qiXdG1)c|lǠ@ ?)lD9H-TH؄_|V| |g>H (E31mL;2GzRU'JpLe j9ɎlY%P*pO4x"a*E)&j8J7~>QQִ(Di1;r}bUC**vI΃YYOO ɻNrd@c.K#0QRd5ٹY0Bĺdrdx}7o 2/0. uյ5ᮺO?X6-XH:TM৤7$?*}sG}5f}&p<96F4Xb_gݑĈxG}mNa}mbǣL~sX&f0cJ:?.oV%ʲ&~fwɪ#oC𦉪rNMizh%ƱHwQoŀdvXVᎉ>:;#9F\dT55TS[ _#C< }hm)+(ʒe||v;@f^!eBTbH.D8G%Q3+ qZB[EV{ثkk.EJX7k{ێ-'MPMef2]A,!ƨDKw0Lh!+%͟i9>r s9q,'U)+ rL%qzϋ a'Z V6,+e?~p}II1->e!0,^eLˋQHEnߨ6/aZ2qH\%qQ^h`Rb\cc>-^ J ֓,͍ﰭͿo//dS^ލxd\; L3%30&=!6ć3AmQ KU؂92V"&1QNGRJ{嬘RX SM”0H$kN){Hn-.y !=zļS GP($7;ɽMST/[y&lxoN=0⤬:|xsmDkaޒ)h TQ(ps;rJ, FyFUBp&$@3tJPwYޜbWrnB Ue|elӓve= ْӿ{P%'=pe9[/y <˛7'!G^s2 ɧߞ8_,|gؔkND>|zr/#{:Gjok܄auRǙ Ytrc(jxy(򏏓 ~8gRS!F8Πt^p!Ǚ!C`WCjCSguwpܟc^iRB9)V!g`! H^nќP.4סu%s}M((U=X%wH+| gqѤ2:8̬_,},,Y gUaX֋X0LNr鍕9Y3I}.Tka<5LC埇dԆ^ON< C <\@TmY9Y~F-?Ofyr zhG -9\6`|w}|c E Zbm0+UibL;ƥ1 \\k94|5eg11\ +A IʻIx(0( <P`Wz.-:r1  Y@b0d-!9<ΔN5A4(eIp]H _xwZRmJ^ -f B?PkL\Z Jq.[YrNVX3[d/aR][#+y9N}I9N^ lrWٹI{e٭V^v*dW/@9l0G"CzvU;q@de8{g=-*oY}Uu#BN1hd0OPa%XQfE:k0H~F#FW`^=TŨg",B5JEpUcOr!0|As@3szPhˈ滧٧AF9V! [O/l=YYZ&W>-#\X33/x ZO?y }( [9N=nzpen{}E%FDU]e XG_X|DXgdi"!9H~;&I(I 䣘Eځ|2@dWyYڡ{vR.(* |Wu<^gm`}Y=ïsԛCK5l󵉗,]pN`/_\0U ANP|pA^KVSwi@q!ݳIRVۅ!e6.{l m;1[|@+9m!Cm"hL SܵP"OTxЊYqwV+F=,?Y$r~u~&T$^ú[77yXGEwUBۗV՞~-њF?_W?>|@o>\t}ت$]@cW`{{5?x;ǟfOދD-2:V .uTQ&A]0yatv%^$Bj4ƌB<ξ6 [Z?٬Hm\B$lFi98\J; _E?׆/I"iЀqrԤRT.͑b1ԟ2pû櫹jߘ@7e"靝}^t 7$a34A~fnq <6NDY3MnRM/9.Hnr}c 8r<~/4hR*uVB8TP;$@1k0U zHԶ`P,{Xa^)e/QQX{|"0+[UtǫU=lOCuOw|Фir,DNFZ}#XP{Yň3[E7C_\nϫ 8Im)8mv]\:guL|DrSm_N0,P#˭4dS@؉_1's#!Oqj_6Yau6Yaud5e\K]PVHa7.0W4(*T!;^RJr14|>CI?ܠ\F+)1ȟg+wvYonzs֛۰6כc)pU-*#CBX x .3J79)J\~(S?~SծK>1sSq(ZIqY2 b)"6&9 -2(6% 4F2 Q*G_52uئDl<3-GQ6`R(SqWT:/#VYY!D : •E$2LCvd#%!jyUpbMU$ A/C9C3oMwtex} +]0 /jRc@bY=2ߕMK('oPM"ޞ[~;kG%ÐO`rܖ0OmJ/)JtW TP8D"-Lh('ә!*XLŶvGgw ^i)^yw_vT)RP"Ԡzor ?|,JX]}撁F@'ԛmrڡ\+~?#R>,V0VJ "t1 f|iNK>F!% *4 DQ}\tq`Gь~9ƪ7ȏ1JYCP_!|:NfS/hR/hbH|Ўܷjk?5ڟOm's36`Wh>Ӄq0rvӨbBHpagp$Ng8W0$1\  Ba+m&g,+S75ǝYtQrdꜮ$; z%L.A87rb(G=/+ר{c  08Fp=cJ^dR)`9 Xr]j?|ԪĊ}m$ؿ7~ (K"Ϸ>(2ҏ#帑䙁Wq& ؼMzl [RG橔G49Ԇ-=sv H"1:ըSƩ x{g+H#L)ertʘkRG lK NP` kTY@@ /7gXlZ{;-V|( a|׈wW#_˯w6b6~G߾O(7ˬ @d}))T\|@dox.J`!ѻn XLvY"!wDO n6XH#2=>؋M 6yliMn&q@-#Lp,٪D"ԜJqd>,dU47SꇒR?f2Ҙ2BK77D׿sU5m{(p^0/L!C!=izvs7-oV 6sշa8뿅wc& bt Jț`CBl?0n,8}XؙwaynP\Sţ+mծ|L  ; {O59Z;(v>>VI1VrK;-o^ ';SJ3C&I Iʾ{.4%y3Omj#RG^ =GӺix2789޳TҸ+BG΀hHHW֔H*,SB֯VӪ$V߳pWhJmwS˹xur@]f7\a[u:ejWIgZG+閜xvYRc)B798<9-1=Q9u(8G,¥6)0UI#Rj2!Sڅ6,2\0$f+cԃ'L.kb/a֖sq5)&$ &2/&@(?ߧ4$߅IGU1ooL^_=t"ZF _ dYHrd7æ|mB!Q X$`Qx)) hŹXKVڲr!ֳsnX#@ ¯0,ϲkoyyƘ3t aK#D֨`sR!&UHNQ{;)I_"p/i'-wۿt@ |zR۝dT$j$Onbh waq+bֿj2p5JhǺ_!Ҝ!<\]|zU_~]#S+Ha%dK8!ޡxtpEnC.*ˑ-~[,[]yGd/43qKopͩo]gtn]ѠVK!YImHdP&Hgy|/ F){c1V)Q:펧j˫uJ_:2caܬ0Iu}$8оYO+bǤJTqx'#)!&(AIv1o =g',k%q9l`Rႁ#~%ЬDA6Ε$Uj\Ƽѷ."<3f<_"(I ݷ@qtQC\).[x0$t10Q0KdB(:wx PB9d97#`Èb^׈@{ԽI"0*mg_j:S"Ɯ}i ;F*awwxO2./?-MϨC/ B㽧1=[^6M֢c߁srt@$KXXpa֐&rK] Q*wSK,D&X/0ØKX.H).QSv6Ȝv[ܮGw;űh$YJ-`Xш^2;nCfm ̎fk8-֕ 0[dȿZ (ʴs3h\dm p_ixB`%[C蛛m}ƵbґbX d-[@+CYdtJ"-.b-s1r\ҳ@1u(poMZӥ\̲ތ4raUe%JYVV`(e\#}R$÷ٰmKO?~)._W}Wj ~&>BM(%Qi֩i奀4"25/[ӯ_1j"&qHlK&/0;tNriөf_#{1AR)~D:=x}yPP==cp=A#92rR%zv1bXnxQb+%/io+IU+~Պ$]Hъ嘴nZ[Is/j4kkP>\ÃNiӃ+p02uJQ&Qyl %3-PK[ ^uP!3n_Cz y:ҭV7pRGh#w9F(tġWYzazEAuKb zxը;i ^tI%=ݻ<#Fө%3s*K] Ƃ (sVےSgeLHW,T%Id[b,̬y IK4Vk[n#GNo4) $Ea׭静.(m]"e}"%oP7DVs8yO8tsL~Zf7kH[/ڊ-ZflQ|YW2^H۫mA@pz`ji2_]k8~Q> %Oο$|swVQ{v 8g>U‰҉E#N|WnX^ayUuXV1⩖rL #Si1JH <NlF‰M?^ۜ{^zbImVbDQM#lJqbQ)IR#DӬ@3bf0E'iI}$0W:kjC# ES1S3L,a<)S9#:!QBYHEIH%yUT Q+J)ݪT 1bUBK"(`K@PJkrc*ʆ݈lS#j"5/4={% oTCFwvEm v,k5T)GU;rZ F3 s hTp0iFֹN-`'BIVa?`B c`k}E9ބ෷r ["Fk%BD"X$FL ݽ+6;cR?[^ [:v*6Zy=ܧ^sunn;oqӽ-1*(bMk[4-Hcb+P?ڔ bH+Qm{ZK a !".ZK쓻NC(TZP~8K0ݯ@LY0tqiPpu!,X;EA%U{ݧ- ZgG3$3bu*ظwM[8˓JJR !WfKj,T͵J)?TSJAPv,Ҡ6\q,3g]Pv%kr\uB-~=Ӭ1*R2 i~H.H^EN!ՌhJ_fٵ/KuA$fc$|iFIjR@' {uKj wI7;9^iiq=;uEԮ I,ZE˗ yUa/ʇV ;kF c"e}2P@C#'K-aLYyZr\uiYS$^{ F&QiKB`Joeu>dtOKtHu}K0M n[֤B qP#@ iz \Hm(kiLJBGRPq\;V?>vqB)-ʈ2g "N}~"a?3+gW8Y2%jJg Lg2Y*rNL#',"71oj$Yr6Ӗ=_gW>~MAܺ$?+/<ߝ354^:Qޕxbk j'=ROK+A Th7Ts)4=fAR+8z]!aESDBrz4Eg']: E Xi(y)ȨPL3&&#9^M*W v!b 6+kZ!}>ŌgA?ާ"~>]Uwi1JW1|vFq㻋J˞\v&c%wgE5@3*I$eϘ$Tsvh&1"scPgvp $ yUoH[۷bAkeA87{nšqWH۞nfTX uۿC'l"JMP;5zo}mF8WD@KODSIo4Xg _9{|K>g-#Of~]"B|U})q"J5 뻏67ŷq>øޘ;pp F -Ut ;:V e7?S =߈r@b??RBX{=O5/?u [O^{/[OR)d;(C˧{gqw3Z*8Ĩ! *6Qf ,bhsCa?W[OVOTP A*A>Э?xܤnM`]zIȲD]pF&H6vAJmdĘДv7o|V՚B5=Սϟ귚QF1rcͭ Pи(U߯Oڭ0R9E rR+Ƨ[-U? (!v{)Wi_˹\ Z PsYW(PzY^OGV.Wĺ2l5'z֥;OZoun0ryQQ^\?sߝϲMkKe1~7e3:_Vt]l^ wAo<>p/̼m:_ijږ*𰷮cͲ,^\qz~)o}RiWIYT\lBDwpbnM11nܕV*]{ޭ 9rmdSX7û5t~#Ļp3QrU)k޵wk~ڰ#7F6U9uz˻ :ivCiEUkwkBD7l =m 'i{.O$?\eZ"ɇM4jvO8H*{5ך?G.7'im6{hrJf2;'\[;OU:U\jl ίs/O0Xӏ7PAgӈJI o $iĬke5⤲tR/;H$V#^4?/m5*qm* 6(;[{WdUӭCld%U{e2TA56j% >޻$m- R&i,R0ҳ5XՏXrA T뤆3Ɣc8I%LjARL.Ɇ*B {JE;Ђzfoߵ[@{1,R|p; 'uwH/3.B\ɘaa0GY>o) uI{t?ox_.~&8rۖ؊-׷k`rkH Z?XvENةʇ֑f"}/c,P9mA-$mɝ#,Bo a_W/VQgmG?E%o| Fm[inM -} =.SJ&R_o?MKjcC)??%[sƳ,E.\Qi43Of0ehYݕ+guWU[Ήwc6ϔ$R*)DDIAȒ< F:/%䏋ѥuP)X/DCN gǰd6Lpͷ_-׼tS_a6}*ǚ%Oο$|sg3^'ofLB& TBUŵΈRȂ 3Ti*O: 4iMn s!W04-'rj9(d$aPh)m Ud\bjd0FdBڨȎfr9n D$K3(5Ҹ;#Z%B T*5H ))&r6RBN^/+14$RvY+ʑ(R4`0M QdyQŨIʕ5?7Y3n'8[COlTxgF!?%Oh:AIDA]uÄ^*6|ioh 60ڪMdٖohG}@ͷ^&qZPA,RiК u=7Ag>>J 6_صXH)s*%~=d9PM; z7^ر]Oݽ(\~u-:k|+&V%wOghY="4=n㝕P%;\ ^"M@*TKf#(Oc^xDؚJH%"C1"h9͛5>O$3?J_?@Bd-Lm7n @w#`lizwo1?/綽n>hy:Hnvxh6YYw쏋 2-\Pk!c\ɚE;K(]GB B jd˨4۵,Uv?[ Sl2zʚ?-h9^Q Q~22b*?fvofɽ]2y|ctՏJ]xln/umkg}:}fcMo˿rm{cџ|Y̍9{7ӏOUcnGJ,Y&θLބC8%X45?=%}xh9W,NT>l j /nGL='w|=8` -kQ:5D1OGi,oFw_87kanuzt+cn i՛[rmh+@Mw "_af8Sbl!c.klHTn|+7ϳʋz*:FN?c6KM/^|ch͆pXˉ Ht[ǃ>{590[5["=ӌP=x@U4;G2 h ϣZ^ה(;o~r;/O;YW* ԳtNٗ_tJ0q:ô` ">Al&=`۫d0GfJ*CS5Umj8e6}>_<>(H`'FN:BGUHh8$XybҕI5$G\"4&'Jx)ӞRcڧӶS̳lHwV!Fz -U픟?C&qGdz^ྦJ9wvc[yT91yjƝՍvlJH#38R&/$lv/6,< (x &"! P IRh@~P$J(%O@cY7_g`tsr?; _G:sؕ1`;V +0Os;]hv|{bTqH xjꌺ"ڡg٨s)^(5G6"q!B_ &hʯXȩ`ltߊ{pI&q?RS3NU LW D؆>\(ߎ!R@vdd|'#aohS8|gG>mΎ8;;OeUblە]{cF!*j4Roí8X*ֆ#IG&bLT-vNjʕ6V56LElH 9&F62+I$hd[s:v_ jk_c <\׃@1 FUjuYJڰU;lͲՆx8Vak&l0~۔ ;2TKƠB1RBiJCJR6+P|P~mx_Hp{omG &]Y6;i]&@T\l*Rewgk eu/%q>~lv5:)A1R 1~lv J<9נ[Eni6g#viwYGT*>4-*〢Bڐ b*5˷s}YqpZz}cnr,<'ۛ>֛Vn0%%Y&WKfVS[G,y6S ֶLfQIt6k |*(bʓ#MK8%5ԠHۅlBckt-2 >﬜!Hv\d !2mRKsbJnBakW͎RcÕ#T|* w[o7|/$PV4)&L=юJҝH ]*mU=M#!JtħiG.lK&,BChIQ Xa:%|R b>5U UPa,c*¤X#bZ@`ʈ y!f6~Pfۅ^*.[W> bN7o1BZBűOa5x삌;+ B=Ɩ؞߽/B>G6Vnܝ<ܶm^C+۳کP]4gL!mF@f(ga)WO̯;ݨ_wzۮ8/zqŹ0=bε|!-Զ[߯T;W㋛ZOCX"/@Owr1*QCJ&{5֣ѷdpCGD>鯑i5THNBx: h v7׃8aX(ՂTbs]^a#a0R YUV׬uc=*FHO8tdCfMz L B'[@«ji t*" !]Ld:=l dV<-K3X,_P~Vp)^XKRK <!X Y!d`wyńBGSCW~`S9B˳ɽΙ?f?dꉙjo}0N2uz[0]EDpuABߑK:t_pO7+,SNV]gDܒ-0 ]QZ8n2:bn=T9gIjB߆w>xP |L'ubB[xTOքNu}<[M4Ȧ p}]Ooz7ޏi(>:{$pg-wkBؔ6pN8lQRD2B}ڐOFV`- 7Wݡx RPC}KTP2= )C(K>k09HS~h@es5 Ĵ?}kPe ćV 6vZޫV.(`[%l d'/r2Bv(m%)Nb1꒍ИpGpZhVi S]^QᶸO;K6DRt ^:r(4A<VP]8tqwzJpu~IEw~4~z͜@!С_J]R/mA8;ŒN*qhPO!T'з*5cAԖ` hZƞ@p_ie 1 ӇlyaBhlx[HڊCg-#s(`}#w. gS aP4n`yɋćMD!kr/pO UX`z -"&"p󡐢 !~Xܖ΍am^˕ֻlf0;H?Lg&9Md8ݐvC ␛`޻H!+"nI @AKAB)F T/?o3/׋][ITb0j^to󋀳?QMX lIijh"t:!,^IhCVuaQKd- h,e.x>q%+Dv9S@(m?m`FŒKPi? Y{&D]ur7<ЪB{ZOPށVt,@_IIN@5N|5H(=N>D`1mV" Xt0u_ty:48 ]yx6 ;j@x6u@ƀV[@'xRJbfD@(dRh-FJ@f 8%PRl^S<9. bg37lo r˳KԐםC3?_&J9/\ë1k̟$탙޾|K8N'ZǮՋh&9Ris7ލn ""߁ogt4a<ݾ/{F#\4t'  Rc|-a $& 1\pc "I$Y(D1D@T}6ź̄n{]NHۏ23Gb^,EbVǀwvf+_O37/ϰB蘇FȆi7jAgօ1\Xs%Lyj! X;+nu&É׏E,7 "? h˲wQ%,KyfMCp.t~fb5'tl|8w㱺u*q_&Z;M._Mfu|z~u~j|W/_~^{ϳHI8E?w؏ҲĜ6o-(Yxou;7Z) < _AbOER>; R5F7Iå]v0@ZPUn#*W6D)ri2G3}0vKspkɒn_IIYunjos\q|藸ou{ckuy777#{?O$;r"rDȇh^3.iz ? 7ip,^4s}\!Lz_ޔ>ң哺NVnݽ= rGB?ǿU"J/c0yՓ~kM畈nV\gW*V\|(M$bA{'Ela&Y&O;.wҜ T :u~<H3|m3o NV;G6`OqLuiսoȤRcR5%8LCӡ"alB:Xo Gx}!8_F+QXr 86@H D* &2a%7%cIMEu(5qͤi2LH,`4ObCM ԟ! \qX' H@ y=]2O@AU ܼ2)5 ̶}  I,9I$Jo(BHlӁM^/{dMex*LU |01bsG`ITP\>MErG„ad;MSʈ-/N8Cʈ.5?E"1/-6*Tl @WTiRO&5u#PsGH|F "c%[#g?Xr#8~myGx\AN%`SYV}dԒG) 6:Yd-]ˢ/]݋n)>Qu}խu-ֶt= UDa3V]p]Gdb)쯮__W#vn+b\`±k]"Xz,`p M"`V]+ ]#Ul#.t(ȏy_XKq@Eb+9+Ns2])~ .XћUw&}qfQWG5cmmh'/ ^m44<) !$?;؉1f8c=IcH=ZY"$>꯺+i§$} W)Eb%߳O J##z`4RW>A a -f}MmyP2h 'PNj9O37ș~2ݏi'̀|Ol>>7fᒡ:9 ?avW@XO&o?/(uL$#oHT#CmהZWWq>h0sljR#Ebƾ5Gվff+My҇*s6JBFzFV&pA?[yʝS9h՚C&P7Vw*JQ7ZBѭQS/ H+gs+6sn> Qg1jᨳb "G>svlQˣ_08;Iy[󵫈ƒ}>8A@o~(-kz:*I: 2ˏƂwӃ֩[ n*ͫh:H#h`% #LULbuDt`4F,PđKlces%s=ZNՎ.;.bauYL]aIpOs~?tqdJW-O F,z*d rV| *㷲!Q.!sJo~L̘ؒĀ& 4ڐD!fQe, 0"WSeGe39fG(BsbmD*P,`JGdZ-E(36@L"ɩ(P Nхf,GҠ+5fb$%! ZZXCgwLeg< pA1Dah&)nr Q`Mc5q% &((5V 7NUVf$bvխxv \j%>[h}O6M gf1+݉;Z^?~Oq\z{a4`t_܃OE^7?_޿A53z·Q0ͿM8b49^pŸ""Geo"'~ǢF)K)?S~4BX &muhcJV⡉he4 !b`Kw55R"&TDliLMT&I,9'*8`k ^dz[`zVƼ`$Xm)EZ<^K4Jb&A ӳV;׈ E~ Z9ǂb-Wo(M4)7v(LNosCP|.f{O4ʭ_"_[(X0f%"H(` /E΅ QSa7.T1Y `S\@nQZ(c@&f*XQZ-j | `άs&f"r.638Xދ :܃} YIUe7z+&<ؼ6 S/wn& TR&?t/; vѤƥxd\s1U*^UinloZ[m2H sο@7'wkd/-}ncDIB["zahE EQ`Cl-K?TdFiwz\@M1aDHS#TC&#(Ցᑑ&6: to1^QT rENre@ZHƮB^@0w[QQeH/( kNUb̴u,0=Igw:aDc#7pRf[,qhjő$q}aAN]uDI qMD#q0" X108B۵Q0 lIlAi8ҵ$7^z W^2}Qxѽ]T=e^݇S>;7J Im |y ی$phrVbJ+&ی"0O}-H:LrKp?Uo3K>| }r!9rZYu֐7ӨB\Mf/ fYYX;?F}wwQ|Lj/:i Y[Bqk);zqn>S^ɨgCǏ1)X]͎4&=BWo#σqlP0Gr =C5IA&~T޵ z:kW7ھjQYY 6%Oڧlӄƪ>IZYUӄ bVqR:ҘP#P*a!Vb(y8OsMYR]qWj [%X"'Uy VM~`FNWƽB*v%SӘY?6i&7_-u݆_ȼ M3Q}o'k5?Q ..l*șheuꐸh֪=C&W ih"`4wAZQޟj ՞TWS+P [˦(#d&ȪgΕqow|EmD݅|^N%EaLW$J}8, W +,x1au71p7.*uA>R݀.]̒G`6oMT-KncJ5Gz>s#kxM%#/we;3ϱ2Wg֙޵q#EЗ5| K#e7I]4'ix_%ʹ^3luK%mcԏXŪ"JKۥG)t_( 8*vƅ|#`:?ڕYiQ)~Bv~B@'0j*{4 -H#B\:?Z! L Ъ$(%.OO6%6sy~g:?]I.ORv~B'+ 1:?EIP(fOh)݈1pVGÄKC9Z Q: {HUzÏDA͛7}on~\4dڗ4XS-NHqJTSDt?727m >&* ,~bJ] Qy? lv_AL[m"q_<@+5vMTc7QT 1S8*F|UG0| K͊A=)~LEΆlaNFx7E4NS׿:v 2Ȅ4$Z͕sŸcչ:+R]vh;FЊ4)T MbĎy8Dԏ&aQE/*ND&o;{MŽ<WV= 섨X= /p!8>vR>Cjfsz'?~xosא^;).ҝ_A t~)GSσؑ:i`XKh@Kw 9G<x,Vwyt&ªҰ;=T2* )9Q Ć9%aZK5 E4)HNYLDNM%X].%uAHomJ2*QzRl@C0k¬BQ3NNDN(bAYHqyIp2DnIBH;ͤ V;3JV\;֐C,_!V4X4G zdBicf>ؔĄJg/m[xaɺ V}YfJxů~>whN&3³Wz~~l4G{SͽRNbJrk8>=890(1[m=O b^ Mt?v=z RKfJVX tw>T ˜ٜo"q505QW(wR 2Y#n8J쾪7Fohvg]Bk<4,ƕ ,V~p!+)\LwaN ϾL pW?c>]o=̍%ؒlRdn:.rߕ[**`)Bk{]\fma2A a)3V–mhKb 3T&*(ș)XD+4 6: T5^4@!2˂pQɅQEnl86w]TPmb0$FprNq N &$#T b5"P `'DA5$z~XƏW4IXuQ#j N#iĺ WsՈy2GX x1/049ˮtiTeGc1IҸ]-84xy}AU}N<%sЖZ"4U&' w9G筷[M-+׽߆QFmp[kK~&}D(5j!OߕclB&BPx푡+#x6<_bG Q6/k K9ޭ Ud;sv͈Vk:]nO'yVk6NMڴOb??ey}Bdß+QܖqSxbqE3NȄq}F4Eu=XxʱU}NqRY2㑒#xsoj˴C{ F+d5TP kPLR•dWíˀr_X*mIɤC6ОwC :i$p*gG)ÍsH(˥+)CDP+U>RXwɌF,LIx.,Q.{4x׹gIBIkBu!\TZv4ssD 6AlKY PnYUvP85;6 fa|{[ڒ_B8r&ϙa4!<9oݦ@DT~[0l7['i5}7oұApxT_Ov{p F 8iQ 3u˥$t_x{h+M6{4If;UMGhp^r C'L3[P 6D4AsJ%ǥ\pO3 ]f*2K8E!-霘'̉Ay[;gߍ$c?> f&RuzI)%6#5=568\ a*lj%@%ݗ sH]9$V2n·a^gѡ3"(˄mf=1cB UrvcUqR3|ك Q,䌶m2%5Wlz3R~mGĢ=lFQ|Gڎtسa VMN>ն|F\‬p[Cݠw >i4npN9 vڈ qMܤn5Yb׷@R`ҁx(^@NeP_XUuJ-hIeb1+[ :C=0RC@|IhohJC0QPE 4hme˳R*Y"\00ye=u~F+3wC9L"f`4 AMSS'(QQhATqe PSCI!!‰Đ&"lPٹa=2/T[C<@> 9:@r3*U?HD47ikd:j+FB`(橀V CuWi!>DaK HD(ZZշ[Zhf*mV)-K /r#<3P9Zp+- eRf;Ot`Vbժeb z pܣkYdc U3B2V5O:KOybv'cxTƪEJ]"ԢFciۀ<1O\nLs1 m5D}Uea:yp4B# hi;%̟kY$*^zY=4ޝB98h)^ްJu'MNd>-XsB'nq6YL^zJYFTWjŽ,/;}Q",x |_Qkˉw>| B w9pttiS)Oؓ+w߄bva,aւ4HO" Fk;Œ*<۵hG@!o8!! Rx?Ro7j5Rdz iG|!`I׹YFiĔ;S6[esׯ S@)mAWD0E`zr(ml}۠xVYOcno5"_hd:L1ʠC=xqvF{M9BvKOa} g"e׸HaΎ=KfԤzɝVYR2* Q=zJ}/&kO, Uk/wO(p z,%Em1q%`hgOn0ꮳ#qfaɯԻMK5~#Ż} T`-?a0һu7,L:sǻ!ruox1w8w/Nn] Ct S2SrNBwMLu|. {2u"^p!9T(Y3gr ן@<>}*11n0>}^..1s :_elڐfO\<|z\8p\C\Bѽqʎs c.(bIzri-[H[9 4t>w}wLa5ksچ42q:dh2QrcYQg&ƣ4JC J|NQ]@+4$]$c \`]B߶OM2P‚UF%l(  rH]ޘw,JG5ewex%ҘdsS$^M-z5b ^wlN+Uh\zoOK|s@*o:dM` ܤ72țyaPlvAy7cT ED8^ *G-ז$`hS>i)eE) TyR۹IjHסF#1yi}%cҥrhodI,u+, [w>H$sA9mTEic31"?U I"iBD\0NlD8fjjF3U\+^Vq-N,cJVpfaF'X1{gc į`6v#C6F _lV/J3wxihbeqerW[C֘)FbQC9!j ?TXjzH ^Bu覬6MRsͱ=8qԐJxNb"-!ז@>!n'xy)#9GLS)zHYmΞ&EVx#0VmdX lL:J^<[O! 9 1͞S6 A/qyE)Ǔ1:7˄.-fJJ!ܲPTM`_1nWB *pضNpC?}C~DܿoKtsHXT  `[6O^yC/Ьv&#....ĐJQ.adyǾ0Ln)u5<( l^WKoa4?@w1 uf]5 ?kuv}x{0KIEOzu bsus YN\hP7} y=LWVa餒TK **[Ƙ[4S; KR' ۱len(~ȹDB3>BC6BJ,/qiV)úi]`u"E;H([(R*TsUxt*!bC.*_J&)D2j3:呠0"SҨX GaO `]U F6G&j!) V>g`΢/Xa٧xX9Tvջ3)4WR'՝]7ߋzM Nl`6ߋLBh." |/vG7hLz"͞NW3-lw4+v)5ߣWZ')IpHI = Ab.`P&՞)ceHL2>jyueqIL9j\|q "|mUJS\|WHgg*&YK ̆ ޛDw\==QWF'62{ gŞzO&}=Wi=Qm. T@|6KoJ3.F$>8tɕNk$.MRv]bC +KmCk8QH#VE&T*9|螺#/ PM!V vI0nr[j0Bk\FJk8)|8A\dk.!)'jz7 ӻe :Hnㅏ`S)nd7ޭqfaJɇ̻MI5A trF ̻Ez.!Țj,a1o8֐FR=JL9m-LWSF=Z8SJ$cI,2ٙ9ݨaD8 *a \0&`e#U稜Sg,I$Øj9'C%).!/jzyx$kAHksVOnbt×bQI47{KX)kWiZʑjǖaP Qj䀥7%(dcUo;@`rD]S,$Wm5R dC" 1!BՈ70=^j›.lr[jѼ5W1SHAϧ.!)'Nz7٘ #z\@')m+T[> aw!oYӏ8w$h[b6.B:\Oݺ@" 3<4:~:u>o,2Ռ/^wQ6p}t w_$;f`p?ր)>}i#?51 ]8il}^..1s %:_qlkufOԠqkr}p5~[_9JfM'bп{\n? ٣U19H%ywߦn ,=6\Ќ"B`EY`Z㲶,4%}qU:/iQ󹽍RvAB I@mmF0zn0nVQDFN92'R\lfaaH7ӡc0Ӏ3!gNVUKY"̘(ne>0Gg\I`[ z؍=RS.$-aJvj؁"[y$ P!lWe % 1VZ9)jWZpR{9|bHj!V[CN5)XIN0Wn9](Z^+Sb ^5'r|G}/Q$,~ 1) @T 0B ',܍H%La iDP6q/s\HAW9,2Wb4uzk\׫G=:ց$®;Zt'')QFH/xDT(B~iJQ:q]`Nk䱟/IRDE*L~γ]߄~;8wt9bJ `uZɪҴ4E),%e EVԘ7MA^ z̎|OAKAMTMOq j-o5br4e%m$`Dk ϋέ_qV6+x֖qdDK&Z[,kgs I%ƕXp\겄ŷ?jMKFn8?ξ_ WG96c$;/^6x[\nڂ6!P~V=o( !e(݃sX:ۮ3xV> 6?a}ܯښ?,yT_3+GXC ?qW/_bp𯰛|@DnzT&Β7G6ڦ@NZX{c;Vf/8Y-t!x))%FҕcmҖb9wNllҥv&g*P[vVvk R7'>EM"w)'cR_nKejRJ!Wz(e+5{G鉣TET͹Bv(pP&ZpY^ia,'Jd nL.58]qdUR *h)\ 6F[֕RTΘp+Nz_P S,m!<ڒGsǤܒ!>i /P R9ԗ[R3̷Qzj(%(=ktš4HQQJqJiMîo(8Ajx?O9m2RƂQ[:QzLmx_O\ơvR.PZ69Qzr(U&9k: *|n3󸴗nwwXrIW;1fu2˜[nAB "ƣHQ!A&DXW•o LARGʟIB- ]M ۢ\84ׂ݈hdi l]| mA bTj6jg#%~ۦmЌx1XH!]b0)r0DBpޗƖUo":YMRVE& fR/eIqY5L"uٻFrWT8Khl~S+ {ծ:ΕƐ)Q¼pHAwx-pn쳕ʝ)JC𥯡Z(<7 RhɗZ/Sg-j V楀 ؁$- L{SL5rC0goڗu4r0Hfʲ1lVՊzTe6lK:ϭӛ[=VZQ:ۨԭT:+S Qg-jٙ~V*eJ /} mi[i]6]+t#8YiڂJ+u#;B[jIw)延ޅPP.Im,~Z~٬flk"ض]r[ߏ?7dM99r>Z5 ??Us~dqk,t▀/9+!)<4JЀȱ!lүukbQSs,&fk`5iQLFۨb  N&| L؄Pڔd21_|^OI$*I4FC+lu扙5"Ct`<`Vf6yŹR G]ەXLF?@[cGbw*p!VSWOuqMz 1FVqr* ZU G OqίHRGnt8}Tphuz \{FMM7'haȕ 7s׺8W~BmmrOVnI`p?p4˝~V@ g7\IW˩ NrscQkLltF3otּawE V F sw*( q "0:(Z8iE3!9 UGW:H%$|YajE[ʉ9E D&\F苴 "2)`y(0CQ^9OњIǢ*w'M-ڠ*>6@&-F]xb$TzIjD$P!rhI0!g?jh9u&iLLu(\yj L\X&EkCHJL WkI:4(\qcb6PJ*hyzEl1l*il/ҎpLUJDF1Q2i,mN86b[4Mi2Zv$+(Hk1@ bjwvbi;B䖦|lBf[^-\mYp-rcd[L(lY'Ke/X렘8Sg.pP~5xpHov6n<[Zh5r`U x1*ɰ{Uhbj0FX)d 7kbiQTRMsh(Um@Xa3u@,bUiq=QNԍV ȱY9qsvX] }u߼tOpJC5]]@+\y5jf5Z:WLЬsŮ֑@-ԄRHhĜ` KzpM}Zh\0Y#=6t&FH?٬ #fy=Ѻ;_7<ILg ) 5κВ9< cvmy͘J||z#zGM4kDfAi1lA(W<(bpg,y@eBynAG5kHcs[Hؐ2ĸZ:{Ǡ\AB}`Q lsݮԮf"0NӔ6KgŹmGRra:tLw4n],z0c`!V7k~l֨r56k`LҬOh5zg,}5z Gk?Eę^+h垴rvXoK}IzieNiZ:2NzK궜Vs4ԆQLk}t BECqЅv=XBs=IM5mUaơIiHYLdIPA r)[eX@ mz@{Ͳg*{ Y[Qps'ga(S`!%41Fw಴@6g2ԕ2ԋ<fJx0MA$)|`Q$ar(Lf^'# , z<'O} JStd桲&Y#Ns¸׆f&A`)t4eV X4`yG_P0iêr=/W_%_n?\ &tdM~^/_O/MqdQBI_>F0)\N/O,*}P$ˋ+ 8W~Zjˋ?VQ/6ʼW+2I"m+,7W9>\(& _Ooj?:,&zjVJ`[N5J+MG5]7VR/(5{$oWVzqN^8.>Q I5WӅU.}_ܲYxwܼnt^v`Zyi)[|uFwƈ*;=[yG(rl|1^'Py5ۤ.L&Ub|&O#~aG[|㸁s{jٵ28C}4*0gޖږ'prSo).{ CwIXAm E)aQ;r^pe7m+i[5.GbXǩX%m|]ݥ9PQZO$.P~(TP.f}7qřڹ2(̙wtw|_^o}Etdk7gKˇk9mFNƷTz),WnlgϽлA>퐫:n++76U`I)Sv۾f-'O8Rqnv8TvaK|8 .u_?i= lj?6vN4LQ}Ԇo묛i=lw4E>A)8(S2s˙Q2*2ZJlh-6?]C˗zwDץZ᪯zΐqHċcWdz^0.}rUOfW%R%7oǁ{gVYhY)k}P\ ]Yo9+_Ay4ibvcvm eI+/Y.̣$ R3.dHj] Gh_ E u- "2(.2GD3'y7x- p3 !ijw2nO6OJ쮤`"a6 1@(qAg pmBܡ wI|.71EhC.c{eؘV 24[ ;'C?]WyK-h.[y:LՂmNh1Zvj C)YW dfرJ%9O`A P;IqEb`LdK+Y?\J4Ո@Ec#&쌽)q{Fx궒D-$QAW嗸E<ׯI\e鐚]- sppv~jTf8`\`jBcg8(I&[SB>594nэ`F`~sw+9\E=Cr& l]/I(HeBϼU6`@LWN̼"NE.P z#ccE ]j@^+VzwkmFfeQ@dD.:R:*%e`wDT.hپlyl %|r(hb4İ#^#_d x3DS'r Tk^zr]blRչtp2F[/h[?=1F['2s-(k(?<9GǤZ=ijwo'럚<9JW7]tե$zw8hL೗՟Sфik==qhQd E#gK}j:vwviՐI9ŻX#|*4Tfy`ӼggM&떺?O3ŨRf8bI*ͽ:&1b#!j'7IuP2tim()E09av_"<*H/_~+ߦ;^Pj,7t 2:fVB Xx[V:kh IdyP&yNЅoQ #'ξ4J_hTJ4{'H"^W suB'D!]dx@fQrB1Ŗp  6*)*J X?w?eKOƋ]3>wط\ӗVb}#_*PDeaOw߿|DBArV׌Y;Flo?ُ˻Qnn n;z;+* }S| &9n{wvanRO38{Ei[s5ս~wy|N# rao<\ x}TKCБ`]Vulq NE1k41Ox!rMt} h%/e }KJ^_=HnJUw}'a2+7G+IH ;Gv_UuVdڳI#`X@ AV]-n!9L}W򯫪nn]bcTj䮧;:2jX! .OƧ&vc%u]Xw7xm7$s4sò@hMq}wSv{3zZN;xbS yܳB6|&ZcS S8V9cJ-͸`*=vN#/oD{Y`BeO pLH 3hdkw"0z@eDRL[!:'hJآZv_떑( t~o{sjM{&^_+}`# ;SDhAR*J Y8νUBޛP%Y8OF%wVJ5V 0!as.͸4pAH@1|ݹ1r[~!Vp7?~rkYe ȨDNnJ JW>|lC0Knws{MFHl:93IN]s-xA]AN~:[}馗nD C4E" -$.HL*$x'&KխY3h#R;pH^DiZRu#jI%T!PpJM~LX2F[` 3Bzy3/Yv32jDcD*R`(!)!5#ARql ZiJH_ !T e$W2j\I4(-DG+ )JLZ-r=T d%ߞՋ0VSz=DEZ9P:Hmѣih-6 iiNH@XAF(R!F-Z02cbl/gE&'Fdȝ3u%kt/r "b1 [ `+yIoM Hz4lE!ttܒMsk'QJ]: m#_Nb5 5U5gXubsU]X yv~Iv3᪹cNpTց-!]=C/iBw\TNjk‡úչɍ:T GD$,8R?][2Ĕ,d>m`i$ǒ^/y-KYXX MK`F8yr1T HH i+YZͰÅqoBK&R H7 iT do<;7 YE7_>MY(ރ&N v&1-QMbKD-jK#p0F&kTA2Nah&Ɓ-VD/ӵYzzǦSڵx$…xG'Ǽ.rN8USU>=ۨJzSNPpqJ&hYrdF/WhlX(zxFKz9PoR|=~zr-&c.l7 Aر"WK5qS7*^ N(CֳE먖bݗ4Bʞs-h?)7/{S-n6y?,=_{n$i6R:G7 k TO{Tk8 DJuWl!L(_-)EBq)'|y7ĞѻʠtjŻP'zէz!,7*N~tkCoOifI([uA61 Tj 8JY%-l Uҟ*Qdftɜjauf]*=0 gZp9\;h9ӫ=(ĔGF-8>zOm Wp9GͤjIcqA7̓U5*ޤMlDN)v}g#z3.#X%}ULEm8JPݨyٻq,WT~靭%Uydjw+5LW@}ktW,AiAR2 |pKnʓD_WR&b 6+m)Xb#%Zb^zE:1"xςu2F~ 7Kvsnu'wܹ*fήv4n.$ ?<\]}o׳w7poI7Yѭy^=\̓Uֿe"SOm}bT.6 _fIX71Г<6k9q ާ9[W,a / Iy #~R Zk 1EVVs-*Y&X֐2Q e6jN;<; :<}B (lZn/p}fҶ/-wWA1Iz9v k=8'&tLFdV(`BF{H\N*P#\! QD͙!*A KlU% B1A";f NAKʞ>B 4ّ1׮Bo#zJ(c"5?=Eb SeoR5E  UqhQIPX) i沪8r>Uv3)K*KNj,Kf'@X+(vJuMjj+)8):u]"O<]65VQAiTHQHh-.ԠwqC}7hBK*Ơ2t̎tE.]Yjv! 0O'=s|u zjRE_5+%J`0 %US EQeo(c= ͼ6|cDXpg~/GH byKc9!CՂ}ڌ晹6c=f>K  B >vWONfs7 grH0٢8’h!?  ]L̗#wݭ>Q`/)6$ tBO@A {ߍ>]L&a4aK>i v%k$Z6Nd-'pP((W~m%pqpE(]pN.M;s펌/cǛE-݈BaƠЄo zРP?#!cҐ{ٹ|}g+>7}'w;<`05B0 \;a;]h%pJˁ!byfiѣ1|0?eQ^S{Cfk>|Ad*} yn"h4ڬx/E9c *:6Fշ0M! Qxq a1Bg lF_n  ֜?ݶϾ']xF̅[q5Ù^ M$ ˛}dL"( Ț͡JI~eV0K~$+.D 8MFVpd:89DD24^BہPy) @$Mb)iϐDp\#]GY)&c%_x. '_,C'pQR̬dR)TuHQ,*%0cHlIusJkNd?.呕3)Rt,?r#;6c6/I2rւC eI .qL4R!YPѕmbZ70$"W aC~{ %'R2ȯւW! !C8H w LP$ 7lB <qƘX xUۤ)AR ]t_.VON`s7a"d[kwF$۶79N{h=`xvTOzmyW^R6~x,ov';Y["Ov[{2N~|񃱇vp$(P@Idɕ<^b3CLD&N;{?S%Ğhllp*a( VSuB7Fmf?:DžM|]{Y=W]4?DSz,$ǾjB9 0~ ή|T#@7_Y"Ba:lDƁCXXF׊|c`eo FP:]w+2B(B$砐b]]ʓ݁;CF_f~]v2oY| ) QcYPk!}| Aʿ\@.{4n/f+= >17>"CC ~P$Pǀz0ɞh@sP)wg BZqZ8n)K>p%͞#R ߐ$'%#ǡFL9=kzpH mcɟN0#o7V!-__ݗ=-VwܳA3K;o.}s?wv_YY(ѫz-mX?1v-wl+ͼv{;XG_׋Xr,^lkX4wrIL>Gb/K_| N`_PrWѣr~o|vS2HAXpzˉ<8CѸ?lA߀ Jy|*36I!Mk9! >_!gb .g3l3AZ`~4[$ՃbA?^\ zI!س&&OYa$=292V V0ٷ;C&gOCQ(ϑ ӊ[f;m^i;ݗzH;,SHjV`uUC:wA]lqoH. Be_ H@oLemFxª9B<[I95v4p^V(B(F rZ9 !G70b|<2o媎9FFA5+JJHAZ" $E;C~瀷sכnT!ݪyxYYnˇ)w710@aD;N1_s+y!qjY81w^xax9/Q9 Oov= }op5f OouJYČg㼋QIWUvSAgKHtn/t{]d /6)L4&D7fK}q|l9!㗋ke6?Y<ڒ|o~rܽ'v݂] *MU.5,)5ʔ k iRaZrf**cÆHjfp|X|y={A{;g/z=[Zվw@ǟ~pߵN~y!%{[;\BW >H±k.c)\AE](DKAEm7\U2j \Pj!İw  0Y*$T$T\ ܛY*K+UDCBJĠԺRԀ!(ջV~aڠ*KjdX"(cS3˴A$ "x0kc'$F=V}1T @!@X T3l!uR@t'tk^3%B*!)SSEaÚB+B# c3O{dԢ}e*9̤#%҅iaAQ6.}Q@ mňb㞅ѽ$|H }/= V%CZhJ$<D?Y  iZ,8k&(vgo=l Ey]Kn7jkh=?(grvLO9A@TiεRRUaX[P@_G!YV F*K-xurIy|mdL}Z"`m9?a~IK "ʧ4|=B0rXq|dr  ()AvDw?/:kpA^rpO/B)4) XS)>\A5#8`d6wF9y$;w\6wcFCAO2qq`9™#+$2 +88^:#!KdL\qQhMͱqv}sw-,]p%A!jTC,%?ˠJWXAWŨⰖI@e5hm[e4T]kWjR`X=fiU,lCM#wxkU6U=<}r6ngߣݺ';iƶ7őn/⶜_VU" F! |5K -%=ɉ@ȣte밒F֍Di>).~*f=(bsi2p[!}0% ,mIZ ?9'Jj:\yLRMky∐26!KA`x{WK#Ȇb>#%B8؞-exD5kSUDs\8Bbj¢ z7! 0bGM=r9)IU/% WWOLjT)A|k :(S⊹(OaRhݍj)0in0Z 55Ԍh оA9DUlͅ3Ziu֛gj_Qӥ62Ud.uuٚ~ $[Jr&WנdH H+x, h4݇Vm_͋nyz}z13\tnC.KO9ڀ{x?< I%[~=yccQHϋɝ>=^ ""A4) Fx&b%;S.U}W[%UT|{twB0x45n׷F+U^4; 0'89{h=!TSη#%)|otoU!U.wvTwTlFꈘC=<[iglM hDmW<gy%?FP鰷*9/R%y> I9r"& ᳈^{"zE,Z# Lf(#Ic숒 *IG2\0EQ c_%,P03I!ћTLf 8A#;XC1 x4|Bq>)dOn!NOZ /Y I{ uejC񔜗5-Z DV#E;ɟm~Ii("HsZ0tDZo[)u=;;+KsԌ0kgI9&T;b(AJet0RZJi ./r/?X fR>HfOpHEXs6c.KiINդLpq{usmko\m˭f)9ϝ B<"E"T32Y3@i@>( Vb/Rb/ߒ(!ꤺ;RBe3:ʸ̄zkG iFS+5JδI/pyjɎ2k.ݑ (0kX^Lq*QS&G(USE,\KfJ+MnʈȅQ ֡'';D 'g,ϐp iF3be`\œ0) Ĕ rTnE$L:L <ό.*eГ`i+ !g0@H:u󋈴9q0ߥP ¯(u9 32crjɫ rF6.ߌGşWcF34OqעI/M~Kﮎ}'fו;^?Рܱnc )v+Ja-yr+8V50C\YRw(*5w3LegcUPN?\~(F,?kZ|`_ QıbR]vzQk]De$T XbgB__)>ty2~4^ ưjnbwvTæ>Kju0c| 6J"۾^ qwhr3P)-ݺ3F2z{&眐VS{dv;qd9 X3ifGOǧ}../1R.o.n?ߍ ̴E8|@Os0=yC*dZ&bx+3+q kԼ"RNIRcEBtcNwjUOvry5^13.=hNliw33?a~;4;߁^U W+,"|AsO\! }1ًpIVo~Xbx|un)zz,_{d+W( RBloxdW$6~A.3`y.ׅ"{ڹ!8W"u vA((V9aQ$R+0aNq!B$wɘ5?`V%"O\aF:2iʁCld4LVFlzD LzPGC?\ ,(#ш$CY7Eڿ!`΅I)mrNDAI`ZVY?ooE+Չ%k)(ъ %Qp#Z7/w5LX e [65vu@(Zu,6_@l{Ml+تV< A__< _raQ=CZqGݛJh >'x43Frӄ-zEOUڢ6tiiyw6 z5n{~z ;h^w]+6B=u.%&nӇ=R͝ &+{!Վ{f 5LTB`gӁϠmyɋ )CT2U.ŎWMgs, S!,Lb'8Z1Լc+R-$K! MԁyX?{\Ͳ 3a6RK'إ245&Hk!Y,A9 vkm-Rp&i?vM&f+ߠ+D%(3CWW#"c;b81u^[rv3/Pu׉_Ϛ-!(e\îPî&H$i?D %oEr- ɉv%3HlKZN=yE gz@nLʃ; AtOM)^y+/\LIfgU͒TV4Q^:mUghT}(w rdgOncl`.;hhNw4.2m [*f#bm"\qVFn #HCDZcҳNRoX3}%O`.͖ļPB6П5o.F}/FO7kF#8}R臼p2xIf?-S$(CAǕ! !BNN- ҩ{3¬ߑ)&Jbҋ˳[׌cw˾\z87QBR ,E^ rjmׯXgٲ&^{o8v_=>_?d@dTG8 H|4_<I_ŻQAP7&x+*>˧>'3|Ͽۗ~Be3B"Zʣs\rRF)(U70gz+FX̬S{Iɛ]\nֻIɛf3 K@,91z H>JSr@y!TRxH!ø'Xd7MOIj0lg5]}]52  A"MUpgV& (``^[+Y[í_ ՈIe[]uCW[3뤿Wem5`cw^/YV^Z q0{-~IpQǜi6 %MC:QKǫfv9W_Hː.W`XSv][YPp\ًVb )"I( pv5kS 4Ng&줯GŸQItI+KS3(W(D'_QoML*,̘Z"S:ѤGZ5F~ Hس]?)#%w"XT((ʑ$<&–I)(>@]}}lV>V((5:e(^U|ǩy&rLjdqBeJ6@2䕈ޕ5q$鿂lSuЃW#&vv3/v08e-Mo 6"ʬĔ0i :E$$M5YR;X(8ĭ7Ԗ:PlC+)7Rғ{ki`a e҅g1)K423zh9˼[fk~vͥRS me5d&mx*̤:JOHSp.=㈱;L8ett-aчﶤDjgե?%z='>\IL4U1yRe?}R8? v6͖p*+ ϥ/ ] _/"'xfA3, CZjR~*[Ґ?fwl -}G6`jݛu˯hА?fG7n٭jK-@NzM=w[J[E{pqcn2qa-}'k=v:T_K&o=u3]],:(&=ЯnR9hFHO EAe[@ej@.#<0Tid#'-X0;4^AϗݯW#`x)LTTtxⴭ %a,ZO +2yƵ,%-*0ΉFQ=1J0mXp'8ᬩMKHw^^qLӻkWe@/9V` `$,X KAU+fSxh;T~Io9u(kkgy4b4$K\8.湶EzI,hxKz\VN[ <&쏉ͬ/p陿Iy1 s$E%"QnZ9zBq)^/61D+3AK-d,*V;Uۮg0N(tMd@8PWwZ1}1A| 7V+q0h]S4`Cx\Ap6 ԤeOkda-#|2/.mSS3նN%Djն\C+!ӆ>m% /\tSgWi]3d7e7ZT3H?EB /]DF0"v)5:p{!O%Mii.a ĘVE%8lձKGNOr,6 w0 >4z/P0#E%RZ{<}|{y5[h;Ljciz1 n|)uv&vL1Y;)5K_yIbzg?җ/?;hOgc&BC{ՠ[DvՋ#hOPCDNTҲF]:ԫt`KRn[O`lf] ^W(Kp$Jx3(JU:z/q%3xdۃO+S?x.] ~? [W JU1[:^j㓫?͞6.cp\`J#4rmZ(R!T* ť"ֺ迩A y,UH F q}$!whdc[,UNjFs!LG9&Tx,U,XT,L+ j]&>Gpg@iE6Ya9 VV~ ^#"X"L c( ;sKNݖԺHp?tO}G?"r?mb64uJ"I\~ʐVVuvOZ(:gwp^bul{ C H-J.}*`_(LNǽ K(& {"(5V!ɍ˒jM=|kKO^AP.{TGX-›l-I f%w9b6 /`U"V X\g9xYŞw<&<ԛZK1\$w[0|¶ mw Q_#P$~eYY͊Um҇{YNG/fEsG#fӽ[nV` LPO9 7+#u(rb ӸKd 1I'5BH^Cd MNQ*2Tx).Lw)#zV+dl肣jNp\YfOk1l1uӢn])?#6&]K }':%J*; ԥ₨>P2֌^Z'J)Zy%,K"C2f"D˒F,)"Gӌ+SnuI׉ 5CLr7 E P>^6N<[D<ŸxvksDn᷋CǷS3)/c?~B\K. A3X)W/{BG/Avr>\yZlL[tpxQ #KJ1\ezA)/=kL#۾Tq Owŷ|i)Oha u}6jydDKYƜ1WDxsTY=<'1x[W N)(a0cUz&i\tOxB sȚ3_E&nKTm1/v9FO~,?se'qYE?Nn8GY3E/nB|s'f6MM΍qݘ[b$aFJӐy]VF|E~B93)AШʎmboqPC+&DfK1AHG}oZ_6e"}բA=|mIЕc#VOµ#s*@{ӒypmS _ -[ܕ\=Wg5Z)g &cmP8,Pw>Oi6|5_eIYW՗LZxQR[Sb;%ؐj鄷'׺]kۑHUI,t Y5b\tY\Ew1\)IkS:d*11Bx8sXH9Lq;dVoa&d~[i! rZ!nW,PƘcF󡢕:m>-og˜# f>(#wƞfV)0cVJR9SRKuwu敁5O4 U=\OZ =AF6>{u&~S._G,)MX 5d_h9y|$}a*-4ǂ*~zW^zߠ?_˛ (kT!Cs(ČAxӆ!4+Z*Z&Sw3zr\Ui_XcH4!qĘc_M#ndGÐ18}x Y^xt0ixqdrL¬~u_BoiѲϮxkQc7HE"}Q{qrV 4%z,oya,AQ0ga~~{=Y^גF'/z=Y:emŒUIYe<]?Tkd{%{qi\:QN+PyAR 1F Ja~A߬ %{k":rSNWL. O7RҬ<)Ss]U,B(ի=A~r gbw Ț5+&?ߜr0lUY:9rRE_?{G+^Zt". {7{ή&M*\مˮYiӒӽRs TaR(52ɐ, 6]UڴԆ@nq}$(uH]@R'&ƘDV y3yOʩQ  !\,TM 1W  0D 6˖c%<:?* pἕjJB[JɎlmtVMA2K4Gv`&-3l ylaCd+%o"m֥%L(=aF౎b"ôhN+=\-{?=~ =Ȼ}&zxۨO𫿃Gܿ9#n|WL~3y|4?,]BL'7ow»P KIs(ղw/SNKG+\yJRyy3*i(c_r%r!c̕ZNkn<׳ UU$w?hkeN1C*uDۆ61ne|㢢]\njf|Dl^Y#V ׽)ݛF5mUmUkVzeKhj(_=[Q]ξHNa Tӿ{/;ƕa k3ln1ZVQ&HO}D&{ |}=87Q=ef?W[쟵-z!*#F߮}-v}wwJ7X"MGEd{$(dYŸ(lԕOW-4kՙ͸ `-cSLY6qg}nӳNq =t~l"+y+95s/YAɂ3mhAAV&JiRXo1~Sʏlә [̓m_kбÖm`S8aCVJW S4-q:E ƴ2&dg6t&9Cjui}wa)w@rϥ5 _4?51Z 5Fp;heTw[| m*yv=J;{ppƲMuI;*d1Us= SYAW%j,rħ r\Vk 3N$C-?8z,F&FVd}L ̑?{ȍMkx,=$$gtbIl6?Ŗln[j3f_Uź`ufU ʷģ1=m:RÄ` mȄ-KJqmRWumhvCq(& Xf;YfRV`;=4##{d]$A!7$/W:8ߘzkԢ3`'Mls> Jh̎Ea;!N6L"ȖwQhq(4)lӠoIItP4͓z{{=ߨ_ΚIdJZZx:/lEa  b!]>C\ vmt s~{E@_jcjM忿qt}|qAqpwGߣ3QX0qfW.z3枘"bbJk5P>fQvޜ{c~V""}CYO)L;&D{zDV֩*x!6\&&J&8䕳h-u馑tA }GvKi)Lҭ~{tkC^9⩒okMQ u꾣t;ejdyݚt(ݚW΢xJSdYRG}4M^jv")30g{{(99~M;KNht'RճS4z;ɝ +VU N/FSDo'vB;ONдܜz;_eOo'vB;k 0 M R Nt_/1J wB]Xm̽ǡ|)fu_ 4 HDD>~7>~iv('? ;XZ ١Dx_Kk ie(~67^EXpz09ݗ&+A-+%'6̽ұ0c fD7нn,T!{qڴPZn\JG b)mї&J?M6"DS[Z>u6=if>9BG[ *D֍[aEw.%l5&V;:c7Zl?o:o?w{;5E:#3̈!54Fx)c|k@N.O[I5mFsc^d3[IL:w<@2F B $KTKcۄu$t9rӎP:Xu7w7kcp|>'8'r_Wa-0}3\ÓH8o{ɚ$Ly# 4 $lBJONx4KQK_J%iOQ,Rt 6\hϋwH13X5eޛ'z}_pɏުA!Q-Ԫ3`PNh)Dfѣjiakobvr(ܨr?;ٛ/fzd,^&=|2›Ż+,3y)A>cf޷Q@HA/~<yQhPXCpz y1~:i ~~p^EUt^EUe6m汦:h͑V2G&>8ohz#:GXac?9Cuu>EpRAVb+a" |ɯq\9( <ӧo4J =ou;7`_￿r<>~umaqʀɍd>cLa2+xKY'IykE(QG-H,`)?m Q{m]>>p9Q*h< _ow 󺻿6q?Q@6r?{o~Yn?Pk ˶1[cҮoԋɿ2c.δ j6S:WTLZڠ8`f)&!ôA2e !=q,0sAf͇Ӌ0EDé9i4#QDƋIZG޷jW R7oW%[6$8 4%" oP΋\IszLٰS RwSW46j,~d'EF Pl4{O UVSSHlW;( 2ZN'Hq/+<qu[x!RpYȁhT\$X~T)|iƅ 8w_nq"Gd V<ᥰ^1!F,o͕w&m.,gq*$ :'c8T:$9RB1 @j@ri,1)< m.8c5e:os[R&)),eYF-gNcAEZ9~>4emY鋏az aqLE4Ҋ `sPA"'rc y;{텬6p*S *F?f!1;/ Y<=ps`Ňx 2D*@=~I.=|Ƌ)(nZ 6WW?f "2DrĿB~|{6t ߉ɧ*jlۃPǁF`f:vxb> 4&Lx |y?Xqa(0&#D8gL<RjV4rj=6-&V G6a lv!|Ow6ڎId'8f(O@&vbtuAI% Dč$_')+M*Nۓ+A>QjԸ"Z r12%AYtR ..竇\r¸]X qc^ե] TI5X ú{.b+H{pIz! V]Ϯɗi>K4BQ?x GVbnzuRns(1(Dubm+7<: 7̡mr woncn8{ԴU68лIIBq؃B6 ΃ 8F]ӗQ1OWWgqXЗӌ7O|4*c zy|6e!i*Vyוk.JQbtY/K _g-d~B4U[:ƮХ Qx!krJ@.-hViTVb`ݫ\Rupmp=TDT%*وmoܧ%Eôqݹ :#71g<+Tl>GnR')Z$D~?dwO<՗@12ZX҅ag`)@˜{nl[R Wvja6;{&ti嶼N!]ɗ@$IFS*i0 ӈ6i&&7[E%[q@e]\37E0a?%EEIIEǦN")i[QaЂfTs2ЌjDZ鶽3<3&pKu {;8^X tKDk7+Ε}_5 Du'@8 B"_LQ&W6iOǂߠEREH&:s}&Rjq`[lvAJUR;WG*}%qx 뤯 6STz[̶+GUT\fP2[EuTc|k()ZO^R^bms4(a1a8o2|W|WeGb|2SA !&cC064H R[ΰX]XlG%@?|(Hp:\N&E/T rPEEC $`}5+9GO LᲝں5}"j`rOnrFRoi"&Pi%)sK A-81"g ׁSlBV bSDk : a {lh39 arP,$6D9J rhT6Nw2hR]XiʔW1y@8xH(N29b(~skҝS;1sa͙F)QXa3cÔDkJK5yU7N`Fx&(E cVrsydJxlz`^Z% - B{s.[(%TOhT80_K+>N撪o8Lپͥ6cݾYZe=Zq6L  dcT Бڀ,'IXjx1$)"1`,25KO8n4.ux K,.&FA$=o3==G+69!HqKUgLH &@g[e#FtlJd흟cOMEqH[Y0["C 4N#P-8,iƕ¬/fM_bo'b jswXl-a7ogښ_Qe7{BK3Tꜳ35}9" У-yuI2},Q abnt7ᷱ^ݍm8L?sc)N](Lyx<^|qvԿB{eX#YZ6_f˶ޝ:_'=fL_&2:xj}l4[ F1`seqVD$K /ݞfwKŜ"]BnK1ph2vw=D;wᲇԟ &fCpbK+ TS}^6OiHlJ#fIgD8 KnT 2PsyoC +ŅE3_WקOߎxZdfs%Jivu:ۼ_UF|=XJ{$nfz~'U#ҍeYZUZ&i uVb>J,,Z3Ӕ"c&y[f_Qwvէ ,ؚ)迮pPTlIuP^!H岁wi.b4 u??ؖ쬗l^År_ E捽g)*/ͪ*Q" mCD⡕JVP!?_B@ M z <[4}ħ6Ggk|0n\J۔ vlykX: L؂0شL؂ (Ɨ [Vs߸Z^yp >.x7{}{gb30UJdʬiG!XQ<1V 8-b)ՉISf`/ j kPa4| )X!_ɐӧׂ`@S;@h0K{0ߌ![vg/u|"Ҕ֐~="ąq -l/A{v^Fm/K1hȉl rÝ*@gĸ`b.Qj|zAi;_f[މfCĖ M5TF5~zlTO>` G߃'~>m$ޯ@٦k 5~s3q:=|w7tyɌߎ#01Wd8Rj`w=9v[60K9wιxsݢ,G9-#okƬ~ y欞GgS:UTH& G)iF 'n_ƷIry pyw习mqaJ;חDO Mr^==׎HOP,ĿSY) O>9OR5#.6#l<~{&5 WN'koLc5};h:3ya_'|9FVA$>N5ZD/NZA`z=)h-Õi16gO guFOpLR<,+i?c.ڽc3n`T5U*F /MKQجzJf0|JHayVXiQ6)BI*L¨4* FH(57X"iQC"_9dk j4ɖolSCTPyGJs%mm N]jMj0qRL)) XpzU[LpAH՚]L8$dvB{hyG 7c-I;I"PU H> UuP0J(캏%]s^mC-zfn>zm3zOTb1y>彖{ + InĊ=VtH V:v*ηbNuԄN`Ů~AѤT'(! )PA՝ `:$UJ[Kf1-rvO;6C >PskeP<(=JeϬj!ok*\ D. pE*-% n*@B=ts@N;j]|+wk~7CQ6=fKN;-ԗ mxx(%bd Na[N fh*殳E Pq]_5Uiz^2ZSfO[TwT|s*.yǐ!ǪEq^-@M]-@i6*'2Bvm펟IozLB Q`9 ukBӝ34Ms3Ʊ(X;FUUΊ!ko:L>] 5(<7# V)X.s"tULX#=':ju꯵iS2_])-"0Bxh kU1m5@MuKP_ `窣+կQ -  j6۲)=4ؚ (pj']|ځ# ]jaaJr,4u?Timq ZՃ.[ ڸRƆ +$Ef58ѱ" h`Pv+LOf]= r4;w(˯t}G /Swnc"_Ox:-y,`x:Ѯd^y2}~2Q~rmOoa/:0DDjc!eP-HefD-#ZpsLB8MZ! ~_b~ >Gs㧟lj4 }Ep\y1^ȝ*:Ywdud~_70sJ]cջGbo6Gut6F.qDŽG<ɐ3M%֜M;`K(tMT#I j\ip%6Fa &aYlc DZ Tadf9f1G4)yy%ZSD:oc4]% IMa7NS ;4Lјke5=_1a#FCp"^2 9u") 0g D[93ggV(R &)bpOcoƋUxF N_eJ'pg9=!ʴM_x`ʮ 0yiz~MMg&@:,|Jt{F%)oKyy>~h?`VXWj}&2MeGKw]Jj]Ó ǻΙys&nucS N zZDLX3v3-5Ҳg" D kqrZ1#H#*rtoE'ߺ_6"'l.{}_ B<)4xoVyNHPwT㽜VԞGr9֨G-Q=RnCP PPnKRPw r|ZvجW5:}yfUébqpRvcʆX=2u-#lMB=ݹ8MFS %xaˍ!:a;cd[a+ba2HbZHvh8>,k魢@tla)]l`(m38f$ B g86:Ie:eIlD\s""IF-1{MK. 9 23e&,385T$eR45 HX B!`o"o4O.[vd:Y|kj7'jC!A[m Z0? !e <*( &eBbI,ɤvLRD 1]Nz, AUԫ 9..^챀VK2xQ-$;YATjE"-Oڶ-G>s0ō XR?]4 ffG,YQB.yhU7bG]]e9Kh:omkY[vY`b-~Hȟ\D˔@D7/0\{RoŒnré%X+$F:%B%K. RK)YJêU![ZHٸw2 !&e"`sK=W13BYDkz\8yl$ :݈Gvxͤ_օ[;JA1Áǜ{'g'!E!S&/r&$͏|$\ ["| J~%(o*Qƞ({ yg|4z0'}BdȬ֙*0wO{;r#EJ;"'!Ժ~bihg_77Q6/}*Sqd(ݷGOw! 7)VxE晴!On1Уv GtBQE#nmOքE4H$|j1h1&W_Ы?] ;+z_+A.,UH&O b* R^r~B+Aty~ɍJּ$ޤG_T%[v͋]*ОI'dD v,Ͱ>>H(G=a/?z¨' rryz z¨'d >_qF=!L] !F0 9gT\ \0 9f)QO88@tVqn_)&Ĺ8*1mU߫r۫W).L.w}F w~nn| AcU:"QZI۬E AC8j׉"m5aA1nG1iـGB\S}2%^vҲq ֥e>TZՙj7bfS<Q&gD=9M/QC,gh~tӝX/Fr @,5vϴgޓ=:1$AuEZ)&H2M9Lݓ/C2oXnt#ܽ<i//-QS%#peU_k9|wg^-nܯ~L N_qiz#N $ <NL AV!"}*&f sk(.Yn|帥D ;u>j~6dofZ[%8 3wRf|SVM~|nYDH KۥB 샗į"Z싊:I4^Gsk( 4'qEwqfSUiB1OOj * `O1=o,IKd]JD֒ ޻)a兔e`%wj` ӷ\3'ă%| yt+ ꔚ@E{.>XeǚMf֔[2O4w<>:+}fUjRzI^]K`zs";/ˀnR] O|0xF;uHhf+.Vж"y"SPB*$7PP`@%'"yBBۈ/+`w|4̋`/2PǤ;Ϲ?&d2%[HEpia9dQ殁J #qQSy'H+˸Q#ss8_)z3q\3"E-9 =:-A<9wm!=//hK8$ 5-A=f'8τG`m R|?@[:`_0냵S. R5 <y՝KIrXsqtDQi P9m_G4)6 (zp23ʂPWsf>Q>axzj7o^}VWSjWQ>)Yf5߼~HvHIjDUFZ7)_`-*7<[;%UMVXPRo6@!&0/#vSF^0s7WM'ʺW2IpS1p{to$0zEw"9FW>DSPI,|j3)Aj)]InQj'C*sI,p53:lp*&2J_& @dv犳|si;</)IZ|ϒ3Z%SV%Cԥb@S]qH\W}D{}EDTIfq Rk]U*UtE%Rk0P\A8WF nӐcP[ȹU@oVo@UWS]RKpK_XI?|o3UR 7~^!OOݻ%ukJR=z+a|[N5u=B"t-pM~+wA;}R΁$ۓ7BHk0y5U_#kF1#P7%cba >߻|sܹ 5 %E n]L0.D7ԛ;V7Z50*5*fA ֚jŐӠƨdTN~%ѪlM>ܜXxu,'s7>dtx~-/?^ݹCW!6mȽt_;MgYȔ ºcV4;btϨ J=QALf $ө*8$=cAA>!9 r~9_uOF"L%@%gd asܻa_9䶖7C;<{9P"h| E+\V>QA\-Tʜ Vwj6UOk} Đ^;O<'] o{i57ERs +"J 0$ 1$կn)O腠t`_ UR'$,_a1Gcw]$R!7F RB p )LUpR `9ŁwWż):P,9hX-%pe-qTPv9L9Lt~-3n?Dĸפڛ!r&-;B~\ xP&#|8 hq8ۘLܾ.X, \xM;iMjv7^[_;0_3NKf08I~~G?D?CtB9>e *n#| e*@B0l|XzJW˜m]MR_z.H)(֥7fڴJ~,I5CI)׭bnd2S)h5ŚGFIJ;* &pOU%F-}'F@2@ Sc I*rں:-h) ~i BNiqbg;XN|'ʒ*GwFWDpkTN{udr¯\JYLW /uM̾G4"˄@];2. ,JE)Nj1%@.k/xOcdD$<_ط?=/aIXWJۥ[3vx}u(zsZ͘l;䮕Sm1 oeuȑ F3*ݠv ̘Ĕm߻j{"eS瓹ו{B +~@%g G]IӕpwN&3wߊj:{ET U(QSZpEjz Op`C-Vbr2sѸQ\\3c̬y}P "rڹ .D6A+s f!_zzCwsQx%SĻL4_<;1On]jۿTvNB24ߞv~JUJاa (]kVbgRm9 ~L2*424,Dʀp!X)c448VqkF|r̰ssz羸hBg+hG ~F-?:s%qh:5h qp `B~=h|XsgyFm\`Ҝo  (xEYFece Νq\tJ*ԯOcRz*JԦ4(P@% 03" z 0'bE4Cfx Rc A@ † UjB*eB|YG# o6 W(Qܔ[qnjCC8#n<ɓT<'+~4%e._f N1;)p,ۑ'A\Ŏ- ;`F/"`b Bx{00R7DkmFol$6{3I)&ٶvecWlrzVk`dbXR5!rqOJqvstI)|;԰)&݄vSNA+7Fj9d1 'BN$iJQ;2VH3eD7N}QRԗO/KV! k\VwJH#eSX&nBX&naqk$pl3΃+DP.HY,sN/EI/K5H3tGGR/-DhzLjÛ;!#=uJc4cU64sɭ@M 4 9HfȿyS DJ[Q6\#JMv2/49Sl1uD0il_b< D-e;vU0c1`fZCxCΊ&ٯ7ӱ J-Bď{n6;fjo f~@פ ܫwO_C(W CoGxG_|_`ޏ'Ӆ} ^\__!SL? F]o}F| pm!0{VEG#4sCK7 sX 2DD828"K AlY"\b!v0TW%HS KBIp,}4!?BPM~vld ҕgA/ZKR~h'KT6\t[M݋Y$,$> "lyEtD"u 5Lltpz,/TR] ig!/LB+q$.8XƬ/i5wE yӻ{ 5L&Zo4|(Z-(5*׹4YkNxA)Y 958FI0D&3GDv1Ŕ)JBDV__wۂ2i'>besYcE*{4IsIfgJu I\tl@al\9o4XRii3͉y0^9eZ!/|0[w⎄g 繴 >rS"Xsqge Q5l,wvB(ifWٛXR[mBP=_ Ç-|\_[؍`''?6)48`y s8H(:t t3n\2HE8-LaI !A)2gGd&)5j=(e?BBI–`ç*mp$Eaj~ᮈ5pnWD.x%{a7ut_i|6日ۛqKV`CQ-ihm ڗW?w׏`NCy c{Ń" ?wwVgn5-!MjSaM7@"j ի |P|/s]CZYʥdERMS(^mi\u0@qNi׍+x2"Bo_Zmvp[C(Xn&:_T'ꅚUJ?yItwA+X} ]6yxt~z[VV->H}ǜ07ap%|L YjY·Kn\X8E_?/z=Eic4^ڽey)К$䕋hLiB;^nFQb#:諸ݎ<-mZWnuH+$1MwCb#:諸ݎwx*dk-=&R!!\DdJ·ִ{L12Q3=Z/T5q4^gZr=\$" *ATj9?%Y~A5,~n- WJԃ4(#R3go'D?;(Vծ@qgFuaG?zc3uQ5/ ii1; ۙٸ#K|k5 fה.1o=c+6<:76 јvk/Ra&Zg@ I42' $qwT{##\I93Sdj\Jo$SBsp2MQ=$a2!3C sȕ%lS Lcm*8.ofJ%R5v*kdjlR/I)r΃e"kaΰ'HgSGeJ*Tب!ّ9%r [tb5d^Z4҉cO*(oӗzT!c.F"\a֌[ ĴVm988 } ꃭmPpɥ`L݀[)3"f*E*nرzunl8 rcyWVhg1CwY$PY?IUvX9E(ő[-;l=%g"z_w8PDXK7xz?$i.y YgJQ ;w*lPk%^] ذ?/`9mPE%(%]){T b^3g+z"c )X݀J,Yp᧔0pshg/$BԴ92zyťĢֳ-)v!f."NCoBcaxW~`/\]ftf2sDq$}7 UP]_\}1 Z*L*fŻ/&1E5O>.mcFe(D+\F< ~1#ZQ'PnX91$Ssg\Qa+!1> ]uAY֌:f\J˘Rd:!=4ˬ*j`18aЎ"TT h H2sc\XOu9 +I4Ny8F&\ gd@0#2@{}l&Em< :יT*ID2$As3AN2jKnSHƩ,y^\EbXj˰'k&-jŰFX8K\'"NvZhNQRTF3Z(&"ZZk\Si9Z 9ϻpSճǛ<_occ=(&~ea%-ܕ?x /JL!ނ _2ijDD~C?_y{tp0oqCML<,4@&@aP_'m\J S".zb}qӠBKV)?in㞠%[)H 5x_6YHpl 7&…_I4c$X>joIEڹ8lPZպcnؼ(U@B>_9ށ^{AEiώ~ uČ3LL1v[+{qێ7WW:ьV\#][۶+y9'[ V>dd),fCzh}?X<9~\}'?vRf6{uZ]ijΜUS gC5y XB &¥,Gzk=_{$f@A)ΆkF'f&t Q6nȧ_6t*,si74f;{ƹsw gV8y4?ۖv " {g1Frcm|Pf`ʻ|;Iyqij&%yY%1nЦ$\I|T2:d ok} ?A"&.M,R0eNmu0^~! .\ud xs`@fg|4Tfk)Z=ZT%|p&39&wM" \H!HebGUz  U-e JxgSapD,Pi@Geur9s^f ںwfιPUr!6h-.6˝8gg[܅jXG{oǧؕXf(nⱛ3817޺`pntjIOAmd~.7Fp?zڻ|v wǙ|\˻ޔ I8ʢ5Y@' a!v*3ˢ[7-qu n|unz|A-,%HVP lUdQDR4hm@[hn'l78r._bѡʖ@Ok9oH):Tqg}JI1hFRJ8]pbfuMc)e)lh+pgJ 6e9rb`ysPȜ"i`8/3u8V&tk| b>8Yκ}(Z??# ^`te3Z21ug헠M6ڬ2.>d_=ЮxG!'-22eiGT܄ Dj9_L#0<œYZՋź>M-Usl - mZHŐh!ޘyBLj9]Bjs:Fnys#;`y7FԐ(ˍE*7BE:!'􂲇;^4B8$.'_çC!#­b.I .0çǤO=Fރm]  `PT)G;`I!"Ⱥ>׮IGVs |,̀SV1 0St*.t i^C SC#ϖj9B%/:Yt.fA!%(4yʹ"D4Q$J!:E)ʤa ^P<YjRoWbM,|p/޶7^]l˄- /px…`ig ̚#ηsF ]?Bu(h]l3@* [O9A Kpuzʉ{*1Nƛ^saޡ0iCQ}Fʔ"d&4W f=ًTx |GT]`T}SY4!߸S1W|+n$?*B6i@6{"&w01cϣFx^L[n΀UtiF6uyJ@/ ZzD\K}Ė=v V \eUxM*<ͫgZiIz&|{.im^}ia@@W:Nvpw'&-=(c1Owӝ 3)kzcճo^# $2ZpR,EׂXŤl4ޡNgȅK'ƻH @0ֈE, M\(M0?ky˵SVA8vsKV)z]菥-mpZ Baꢮm)Do 0^N3# !cI:Q _Cz*T0z:ln0؅i p$ "#JZSA`# N\&9 )M\'1ւMՐ^szy =<N,NL; 0Nw O,/L 5-mMye9^ov'p!2]M8~kVGq'[V,ա P ;V|>|?gY4vD\-HBވbCMB:PNȾxHa .d$B:ӄaKDAE eC}قA22i`,%B*\@7 $#Vˈ"DHRx6us4$" ByF/ YfxC)40("gJ .~EY)F`=1Y<퟼[&x}Y8lpPgw ϣhMɞL3-Vl8ND$ (l/}bX JGF" 1sݵvd/ o畦#{D=+bˆ1cCzMq-qw!'B@^u|-޾,WXIMDh)W3_l2_a4L IEk #pI$ 1Ŋe"B Os5Ă`/1zHD;.J(R?`KD@%'y :eȂdq*Kr\$Dn.@Zt:o5:2;'YPs"Ѫ\P_mevۂ̯ >J:~|hK_h 6m"}=$/E{P_Og*XF(},>Lq\w]% #|)& Wg܅1iz96Wn b nMA|5*Ѧͺ聼ܓ@4mzDD1Նd Ic 5K' Mb@xrՒP'q^}VM/'3g}{߸FWP3柆өfǛH9I}yՓ`[DYugE`#V~IS1>uhV/[#Lh|avdwRJRQ2"0 YlXR$woVuĄ4@K{*9s)>~UZN)ͶՃf4^<Fn;]NnKLeÃS f(R+aW_AD KG O ^.63PSF#T$bCT mb%5)DK/}Zhd6-y?Jˡ>xlk 5(ʬb%$r3ݢ @R$<cELx 112c `V)R.1O䘫H)/9"]%)3Ui`U#V8a)3 @$Qi0h9.kXI9J@aX+q# ujRH)@!G*_Q&Ta4VIJ0=T@İN%.)giYrB}HUaG)`ۦi:O&>L2*d">#&֟j7sR*ƈnb42M0Đ9pޖ2@XRM}hm^$(j8w03~t>Ǯ>y!x,^0+g,(합[h:s˓Y#{)Qy/N>M߯V mi\vvGl9 qu/ͪsq/x& j3p;z hwPu(?;tUٻF$W>˴/k~gDiEbFxYEF-5YEFFFƑ JAJ/ٍOwn2jWQ5:r=Ww@ Q/%h '\^㰶]w.*:Rbt*p'6E/0p; "Dܖ'Mgj{PMi棸fZTke]r֙TEk8%ϐ})%g},~{LtdǍzl$qAT(4%.S71B޾|9N㌬qA^Il|zxt8qzc;WuOT^quQwtNoK9_>=Of9t?~h0Vxχ|N2||.=\aE9,k t{;|~0cRްo&+RLg-C^O] Jp !`\ Ki2Ҽv qx?XOILP s(>fݝJD@LjțLH -3|"U"(n_1ddV UfwTYLE?eV.K$[;πcQ;Ѱ , jU:EuilK=ڱrpN5kR$FKn 9Ftu}a;^ VIJ툑+A(ClN;,ҕ~[f5TCs w<䡴=hyhQk&@ _IrEApPNn${u|eIrޘ-pm?!;XEm W!Y; =x`fUpkeӭb* .S ( 1ht2*ul^ҚENQ՛ıbw*3.mgesJ':+Ls8Hj~y^"M5*]81?V%`E/_߇W 8y<Iގ]XZ0r8I?QZwFiݍn9,EēW.5qt]ϐ0XodNb ~Nq;wE8kA8[0==X0,˿nBlc_3_z6uEc6c#boo417:0!XAǁx\8ӰZDEhn>Z*uy:d!B笂qq6Ȍj`FU`9mH4c0cJ[` rbDJfc%ORih5 fT2A,c4F ჳZ8IC+hFڕ{`eҬ:RES+`vE{WI&Wɕ\A9zDVd>=7987/PSxz7UPVbkzw=*gwvVKJiF , XHX !V|Fॅo$b*1SZYWka$ʮ:]|@ ÔR~>wOzc13@O;;{N^c mFk:^ie` Ui̳{Φ]R:[a|]ڈ.֩7 iJ, jZvm\g%JcZq/hf3BY\`鄧rӫpZ69ݝLu|T/Y~@ 򾢇VS5koog9yv2G3}9x)Cs1`zf5%%vnOwn2J(\j]n[D{h@E)8_V'U'T8 3;窥sWb ~i,6By} &O #gk\TbM^3i'}5.g Ks)bѓӄRfA[iC/aLui-D܏0\lۏ9q+Rӷl:dbnoخ}L4 -69 *STtʄ]D2KS|(_q%t|?)s"G2ÝnML$CW$eZa" kI: }~݄)K3x&tsua5غZ%p9( ,xAotRPEr8S܎Z*&nbݹPcPҵcAv3(3x ʮ6 t`pՋDp@ *4Fc*q lI- ZY-^ao5Vޜm`Ő|_nW>DZ QEt`B mT8`Ɉ)cOQMe_+mU 4M1~];=jznꬅUM}늑~] p[9kd5T6.mnK,:\4+BAy4vk笜dЩuKXCdyd;^D/ 0h"d3, 2|iwpCK=]U*őK7xcD aC*BE\ls8yy7tX!0/LRr~;/q sYzeǗ[2D^ܤ\>B (>7 ӽY(o;o'Y {n:#Z 'l\*a&v7};/[>nsU9<-UqB.'//'Ì3 $Br[ jࣘnGU8BTH)κ8J!ˊ?S*%JO߫9`F)`F`rn-q'uGFÆF9|CT*"x$~5&x)kr(Ww]GKiu.Ub=QJTj1J(dxʆGy Z-Q{Y<TQhb1PARd/|C>HT+*٘cKכsX9Mv)n #pdHIb^'8&|mL^3ef 1ꂙ ;g9qY f-^(?') `$HL60с2 1j2Psg f>p~b+gVX$Q8cpOQn'ZRW9w51PYœ5aWajeN&&Oa*dK%e7?ۥxI/uԉpkpdmP: Ea(`H*^ {) LD S(eMҠS?d3hʴLVPۂq(DZ*\"K~$k{xzhhR{Es裝Lp31sHI)?>>,=0w " D k !Ώnop§ n[3}77߹__O <E)JT&+ZK=\>2tİ)vVhO#K%G8{eW#C0#)RxܔT]V|5(sR LSx,c Sz~cL^S]p*,?+>T(ιzS~x+F!pU(2iW|c6lP {M=)+SwKj{Kʵޟ ƖH`M0f!DSX]G < 5XST3T#8V!pBV c0u zWOùYpn$w 82ONfzp'1E ?1\'/[qzsw~R|2|&ӛqO-.p"簵˜PMaM'L*էg9;87/PeK7μ`|N3 ,M0Z 3 hx22qbC2GWWLfA$w8H !L=-GOc$V](ȉho 'q2 ,I?*G]ֻ0Maۻ (&,~|h6Za`Rpͫc3 0׌8M^FQl#UJZh|ac`A [=KC0?`y>}w70-Z[aDh,`J-:jE 8&br4ݻ0uTRa>kj#Β尺.G&zfEyo=6*oQ*eݍD$81Br3$%N( ќIcR?Sz0{TG.3"ǥkqF0_ri~l'vmI>ya6zA;C+4RKj(ՊFb_UŪh^' G7|Mg5F=eP?ܽ 3g_X~g4r;`+,}֓F/C6ޥy߳6G P#ޟEAĥ kHfī2q)g[m_%JK,D&uxHS.k1"KJdbՎ4Uv67:D3Q:a-P3h%fňq s]Zh$O2zC(56!=Z]!y.yC*Ew6㬵 bsYA\Ђyjv0~Bm$nI/y /[1Z ' $J[ak].B9"G)WHo.|?qvI_!yX^hwн?Z~,{j hC-!9n֪AjmZ4NuSm [7NiŁH488jEA PhUU'&xL.tZ/.:zEk}{,߾Z!K#gp 4r*XXsCOϟ5e8ޝ5r=k p3M|yl:vVN_KgL'vYwKg>Zqiъ)N^'}Y[x \f3W!8Ǫ.} T^Pb:c".wli*n @ҭ Y44wI7A ]Pb:c".dJјtϣ nUp7΢1p ٚ^jb~w0XvXRa`O̝l`NU ӂo 7 b/|{6f6?y^BB:Hp^M W5H ޥHo TRm,R_RQ/O.RR*"!)#NUghcoӯEj&^B 1npcEU8VJ E.,!kd2mu1B+8DP!3doY[g{G,G(vgzw_\Oi"13֔48blĬYOh;A:l5]B] BP/MWk˰jFDw@]%1릉5Zef6-O.BTH0O_T5A4xvUP!:8?27lA<[:Ahs >'E), 3W Zcj!AT()tAl5nPi:)Y$LU-~)hVz~îDMߔ#cp]4@ ]ǘMAҚ);H+T sj7qjWXy4[{7eR^&/vjo{,ԉ߬cb@EtU ern ٪{lA,*=nt-#W < ֒@48mCTa.d>E4*qɞx.%|kaۙ}<3?_wɾrt~/6*? 2_4ٗS3꿍My(qu:SN] FK>~I >ywROSe3AlטLw6NtI&=b8݊wl6zȠwgsbS`.I0ɨ6c:FhI2 ©!H[,6~`K˃A3B.w%sS+Wb ׏),{}̛Xi֝tN;'yQꔮM^zE Ɨ׉O?~'w.ݢ}t~%Ht6_1`Ң0P;wOw݂`e kr`fVWWX>^җ'eޝu+|-F#˿XY@3Hӳ*2moXZW(]72#_k309:fq0b?TP?#f)Ox1?A??_;hǣ +~1'q&`vޜ}g.#dJQ&|E)sE7iht>;i$ n$;BBJ+ |pn㮻p쳺ˆg꪿HQ=fOa$'䠊vwU0A4]&L2MqJ1RQcSX! !6! `*h0Nl~fP9eSPbd==T)ʺsPhFS r(﫦-WD0릆mwTJƯ*I9][ˌC4ͬq.UA3I"ٟ6%}N)j$@B$E€lȤL )(`$A$!M8kYYqyYBxYw;->SZz35 L{_G";,Xک"`+T`jr_Ma1yԙ/fβPS~-o~ezWUr߇6U=HO_LreڟsyIc3O\8 ?@*Txr0tZL8@gAH $Yޓ3ݲޟN!k54=vovv؅6tOn*.hcО,A}ю(ƏfweebӃdCQLnM[ >|FE\{4{pJf59z{B`'o!ٙ+^z/l7 PӖý8O.WxzLh5i4v5jvx쏖_xŻ;ˇ짿ZƐ2~,0Z{'ƊS̜ilC:#_Y5)ޚ'C̫SqY>ȪiF!/^qYc=#݈S Ko5J@3hi]&b;|tQl[V~P'џׯ^OBlgG#Z6A㍉tL-WI;B>_ ?B[7/K2bԐL}䌓&OZ6D{?HòlMS"q&= ;&\h/0/kÄkx*?g٢󐛋]c!Oac-%An'KP€@=j#Ҍ)I6>0F!BP@؅s+:q%HOڵ:+psޘw7fٺq-F)i@͊L3w319n :hhWTa.Kw9UFQr=VL8F]8u%5N; 'p4Тjy밷jyqъy]/*c7 ppz^4̞;s3lK_W dNz|~d@e0=娜·_=#ƔPi9ޡi%YZo~)?mfמ ;l\ǐ+U3+Hngfl@Z01>ɚ|oēbeydrꃉp%O>W}@ ws.UZqۇ3@:޾L 6 :i(YaޓF_"hQ*8  Hڑzr'5SuseKD5br~#]v h5 嬀)Y8b&ܛ1vLA5 DfgKH1_xq(B5x嚘GW2dz6ΏF%ł.7BO?Ǖ$5-qd*]x2N*uAryH^e͸l/-0Ue($f tP>T} ܍5L8bky` Z-զFJk}đ22J0v96%F&z-Ps ӗ܅s)|J S LϚ#/ үΣ1ßs6Bؘ0i FC$bl>k~@+w{k׾(9NT~3)sJ2PkOr e|>Rf޽Բ .kΏXx)0!U{LlٗO?x֋m% M\'%c0U) h'E|v:ݵ\URhbMv<,ZgH_Um[/^S$2m&[||J̎6N*n鐡av& Q4Ŋ,IX:7ƪ(FL8y}qTTd_~ )/~`׬!ϛV iD8~*ْ'@M͎Lnۗz;8 A9ȋwG@V]35L[O\kjW;ى8kl57V)χqRJE%OV%qOoWf7ӑG'GEc6߲un]'gJ)^?!@4E99 {:~][;{?.{>u-FY֌y1UH׌ْى6W}Vsqm5ADh~<] ۈZ;gwLY+4!!Ep|pW5ƛ4NПC;R? ս5.Q8M] Tɴ6c3xݻl&r yJt/syC (_w%fz>o8= qJs!@nnuKRAUKΫsw… :USrOA)!V4J;dO\ը?[sj_bv7w9DJ0B1§Qs"N,mvwIg3) $ keKnR ޸\+d*$Ϲ>\La1_Ng6f 'wyw~xC<$5J&B"AVLj^Ӕ&1V)":%#% b*w+2v>AuogpRU*]Di>&Q,R!"cdʤ~QTje*)²T ԟJ"M !.QUT:#/Bos郏.ެ.ެ7kB!XEcJMPR*E: A8c(ƯP%_G0fDSV Xӏt۲tb. hf&0\!,)3e:mҽ %Նo2\ēY%ahV'S; .Vj-ڰ\mߓO ʏ*'-NH"Vx͒/w%¨S#D4s>9SATRtKEgHc\}37p4Q/yY7!hģX`.K+ELM%┒ >-y$ˉWuQ 8nrFր{Auob l@:ɏ!g%U3Θ H46sX!0=8UCg*1OaKX(2Tf#Y-c53mI]A?XBvy0u5I$!ڄ*AiHΕvB5 DPvB092%(IX36 Wc2`jkTLG V0trw_6t ?*Yai+~|>n7nm!7Dn?x:HVtÛ?&@$OߐOrLW i`A1< W'*;zՃ?3`xw݃oq7@aotŴw!<ܔʰb:x;" }'^*aQB6U87a!:nb8HZf* #&AG,kM5aQ=(ФQBqc2 )̭GL*!)LIhOM(v8kyKͰԭ)ܝPpsB{(_–2Ƌ$GN)_ñ襌—҈<oX{@=~8,/Z-f )=~8TRDUtPf4ovA9k~tswǧN ~Hp6pgsGf؞sס%6? G_Ab1i0Y>d,ػt{'3 mLaNc`S2z]O[ V3ōq̵aY؈'n `$28N- :e4^&v ktpVw??O39` .;1 vgb:H9љCr534LuϡnCp{2 6Dmo9"*Lc PCţTa%#&7/z'3EL5!+(n'y8f|e:BV4Kqs |JV*BB`k$5QmqTc< E9;(,ߜ؉ lȄ!IDbhR"#JbAvD7[R+L/ X<yƁpg`|d3h1Ueׇ^FʲzH{0G7!|E,G74u:S8JӀ?)xw,["x]l\~}I鯆Ssh7rCcR202V$%ۚ;w4F>ۙ"Lk!;O'H=[0ׅb]:=k}r`eމA%l5Ía p%v|^i5oob1֍a='V|493f%|@F!'a|r?."MI ` M aS5X'k&0>x8ycϩٴ[{F{\oa??K3H~6X%]bnLd֙zj}]b(R֙7>[Z6VFNꊟnb1sJrBڶ vkg_r=8#M֦ i[-ڦ>%Z NFvsXl$2 {niZ;9h OН_k|o#|Zd{F288hY}0iQB;FQЬXh%$hyR @pOPD02,L2aVZ!D$#0)hH$44< q$di*FP#y̓At0݈ܻ*ӹ] (jO@:H1SKXI H"mb6*m x<\ ,Cf4l _@\+]@*l2s.֣wk7qȎw~\] 4"Lʒ Ь;oѦk3 }32 |oiiK.ʎS,h87p)*QoDW(k=5|^[{|;nL7v=^"[xot[AQB[vÛWM$m'{B9oM)`tL?:bTQXj\/gTU AISsq<}3QL=̣U9p$H;e3Nnn(mA.p1ޖ駸a3so}Vqd^.? \4}c/ wuZWXL}~W1"l@ skYT45IFlMZ'O9l@#Ud!9.n8@W5[Dods6ku \? K6pusLXpcU5ܴS`q(„w# ٔܡb.c Vy#/7ՃLE7E~Ld'׮/All 3e~R}N!B ddF wĉH|g($1"X+SE &R$% J/?W诫`^@8 jCWm/`fz8qnSͲ6XrfI՚ppg)]ʭ[ Ly'z`H4VDIJe*O N{YLYXVkd5TTP殈d+wm͍Xɸ_R=^;Ul|\=iݪu/Y,8xw3t>ׇAd{7o6ᣯ!蟁Oooܤ1W|\Fnzs>p_3R_Wֆ~gb yB 6KL[y2#˜8EըH?X5?^za߃anH:hQv&lGk/@)mDR&?o_TJQ;N*9hf'!(FK@b NyQAo) G}K S՚ %Wk#w}QW "]8|z44h:01<5D[̦^0/KL >/'"OQ,FV#,vsFFt.B] gven^Iy*7Ws:(d]Kyն4NX|V'r5OޙDWk'0!2-GSXscskιN;UInͿyWjp)Jv] q׍iT IYgRf_Bb`ù׃Ώ@TAP{   ֲGWQ7ֆGb@}H޿Kr1O}p';?1'|Rl*H1ssf}j&0P ``_Y JEb%bc_ \B(bqvgZEV L"kTX )Of)l< &PfpVQU\#(mp"`KR`xnSҠB n lRrUŀN 0ycC9R[e&j(]QN_g[ҩ3pڀ޴aFznL؞{pR=p+ݔɸ[(H$ DIBPN扣NyB'ުT`hB%:I J jpʵ2E9NLH$LD7ˎ~(``T$R`4Nv 45hU/<@:55Yl… \q-r> #nF. C"G"=4$wtrܚw`؄qX3wbBXAivUwׁaL>?Yd&Wxn)h5nq<}ꭃ~8^7-0Ee =4=uܥ|nb11F^F9@7h-Yfa,K<%RHi[xio>2M.Q(%paIpTf/x6gqD84mڦ_{,%Q+˫7ܝ{IL}1e؏j%h P™MO:]Mv; YMA .F19v9hd릲..r3mnmL`af M}L?%sQ@,|7 eC@#@X^QpVc[AA#~QI~E(6,eL+Sp?sP}rL,a$`Z2,M">̌&1>H}<۵u!Kͱ[;N:ZUOZH b+H*$ېG ߕifԹmwfH|NϤJ*2s fk7>,Soz3[n343w3YY_m̨/|~7(wKCNjg}RrߡÓerlޙ9qrvYKvܗek;F1wrϗ# kwF+KIɏvrm^rҩz~ : bKXP z߭p@\K3[E. .v %y b7Qs趇 BKw6kY-p|5-AD|W;"\t 1®U~u7G@@\_j L+^lC!3U^,C'_$ջU(3p5( h%#NymGRN1c*w^*yU˫bdHHɔJR[RYa*T[HQ RdR#b0 .a/dlI!RH+ `& Rj!J0xb4@iP5=Z!ԒJbI/XHn=^JgIw_^fBӵgW?;@JHRv(^Ii7LEnAY$!g[cLs-nꐲ#(I޵q5 ]^I`}k{A'wvR{5Zvu sp@)hµsJIerচBʚabFpfB)TlVl/wVcݱYSP T` r<)J y \ @ƥdT6 X\oLJE_@KCU*ArKTpkPz-CoԣDfm' D?b` \&lC/6.$2ڧbkSʛO32A>ٙ{҇d1Z?`/7o)A>(p]N\}~P!wtrܔ[LJ2 ]8=\IliŜ]=j5a1h>T4Tub%[E5.PݼP`؂w-R֦Elm^}m%:UTfGSYU&ap|uQeLrѩ1k =ZPZɛƭX(EDl;Ed-P&b 1Tt“R$ l1$o9QF@j!CAeXD4 Ojő,LV]s(dԜ9(.K<ō( 4>}sc`J~1ػeWE*xMO[5yw[źON77jz;իo}lPzݧwbKg . EG l^h4JN Npf8VIRLəePQaR$bX$lչRltlԄ_hRsUQXvdlNEm"|J|iRGORי&hKݻ6t|mt<-X/# kl?oL©VXu"fN77*ⅽ{ezx{]~C=29j #:< 4.Z ٩xI8比mP@)8E ЕI]w0?3SE,˺zy0%vjbSt-M>vKqݺ?)!1bE-;[]O^/7)b>gTPJrMrYi.kA81EeTg,=ID`1yBj oMݿ *-S"WʀM:EL(='ŭ5΂ȏQ|ySaC2y_"ڗlxKkdidQW|W?{Axpsc_̐*ŠJFR78flCEuA}pk_+籠3UоbDVU$Pz$2QV[2ϟξ/,5#5ƏO]|oWss#oO~Mi+{5kb@MkfBSM#c*43ys's,95(I1!QՑKX!3.X̘UȌ|hzCي)C$'"@rj"ysLG\ HbBwņ88XZMBa jCl-p),B/0z+VHc hKwDPͤ#ԵHP0*ͭ#9.@7`S/Z+H`IP fZxgv*bpXK>r19TXZ jW" z'z9H"F\QQXXcZHˍ738^zId^1L]Qa'3-KCNiQ_3\٣^/ W__5K1QMp1øu}\-LzS( i{ӳa84k^8|j[$7\9l+ss U?!g\̊bzw`&D-=zvOar16_OKA6=hw<}(!}1{[+Ln<J%45)gFoݠXtW"sqIen:3ȳ`lR) rkAР('<׌+cMQ]mdIQ(6$̖kyi~MpQ f<9 ·eV7 1h%mQ-&e_].oOsB>nY*VSXU9e:c05^dKOÓ6;9ޫ#<1/7B8K|6RX~]J p:-pߖlvs߮f󕇿K0a:WT)JI.2-~Rfzbf~WŠcVmxRŵ=N]B/W-R*a{Lꎌ|; _G4! !O5Om3fQ_zŭ(S6Wzre!O .xx,wVQ'`5T\kZ3!*X+}҄3pY>?/ 8} h\O p9nAN8Ci̚b ʐ`xm ';l GR)Qp7J8#A,+SYYZ ^x'++{֝ZR%^ l-шX Gv&VKD R0 '9#>Rk,ם@#9jo"ZMx4xwTefʲ. T. J5,͡%e%t=,{FK?bW*l溼αՍa$%zC9;5XFXz}ct_2q==IH70f_\r,ahsб%DQȝd,SHS֯t؎߹yXR19[P 9YZ-L-?)5JDS/ M粠 #XN󾰠C; -+Er€b|xζޛśϷG>bޭQHzf_]v{Ƿ^]?|߮?:0ŋ??<~'SCUX<Ͱ-b]l*ecRA[T\q7GlytpFpk$:e!Ƃa2-n #{~f=|hn1*䴸oZͳs}Xn2P7~XvbNl,ۘ!*m)-]_n=(cʐK'E!MS %)PU O0xlHq> Wn4ωֲQ6`{8=ՁIGy:#n86RUp7GCԥJ8hI:*#x8ʩE\u8} c!zDq T>%d^KE8}8FN .lbaM0uT Hx.$bs!Wh" V7"ros9dM47.ptW8%'*a$e^q 6^oH d.$kŤ&gӂ:^BX)%#%<^L8HMDKB矞!5R*Q\KN};sN^˝6~5j&0FUhA9y ͉W:nrĭǜQωSVjK" ՚Rɹ˖{PrQTDl¿ `n*tB@LJ7TCV5R-¸L7mH"A; zN2,/)pLe([MܩxmF ![͋+äGqO{%%eW,G#C9'gSZ$w=0D_O\As1#*>dRGM>fÚ h@OUE4z,u"M7睰kJfmPOj 5'^#!1GpCzYޭ#YOk1CTD=mཾ9Tpb@>)C^й]7ssݩ@>iE֑(+NZm/VƝ0>T8H*=V#i~<V#5`f`yq<Y<%6cq#p K@IMMܜz~p6=&RRۻI)g`=]`7(| kn;~0Tznyxp.R #E$#%˜PA8Ն[p9@{>oheRd+T^A (AbuDv<.ф[ZS2-nuPiHS)a/#3LcnM{C+\N=zB?[d67ULrv?!Æ-yk+A;& ĉ"{p+{i#*<|h#ޣ6ӝ& R^AӃ3 8*\|rG-L߮% כt6^@bɜTyƔ'';N\\~ۛlxV Wa/ˏ\+ΰ:ʽM)Ϫ|Eo^2r{N*}-E),O~Uar-~H\D˔fIGē1dM'H$j9D[}9j:ERT3%[Tv>س>ޘ[iЄa`[y_ᓭ'Mn)Q6EꊉEtxoJK͆Ͼy{f);:T;`O\_zlʖP"c>KxŮ% cQVi_lǚRâiJ-($T.iʚ_ NdE}Rێ/cͥIBΤjG$H1_ڑ_Ên7qIlj 64'}ҩ 4kǍYMnwILe&5 "lJ{JM$>ѱ7'$ճ*-~rrVñgLvnnOD4T#a޵5ffh/Cb{O+^g_60͵D:RҐ"E ¡4+^#LFnb*݋nj΂VjTvtƽZ')7'Gu_5'jhO\}>qNq8)|nu`I&GAKsw~EƸlgtjD )}gr 7WGv"F.І2Qb9FTRQo?y~/emw1*HfмFL 2s%wӨt'=NJL? k[2|_& 9IO"b;NɻG^ JTKӾ84AyO>< &hѓHR4iPCL{+BpxARS>V> VyW{P-{`1PѴZm v[}07[R> {+8.L Gt=3ԠAfk0ƒ0WDxKh9MC;7 5 ]!*V~eܜRiȰ/ bj݊\%-]lʉ_,"fxsMLFq{3|);c(frzc_^ff.'կbl[RiWwҩeƽ'8D=\RnPE22DcMBX klAǻ. R&ڎwF >_jNfҔ~;H#8S l$\)r 3 /`1_V ,aa]xҘ`rtͰWb[cAXr%R*:w|. paP$p 28AG.9T =®0*qBO%Hl:O9S.ˌa2;gOnӝOi~%K8AΧvN͑&$Xa"3"g1 G 9ʙ!7.<:iܮ@3l6hbǵ@d.湤9&IjSv+nΐuS9@(ذqJS#%"xc!>ZcuKQF ب@Z~(qr,#[O `zrdvºiyJugc}-#X&x0b b,v^1O T2 XZo\&Ay;BUQ\MRs:Q& !sl Q+Sè+Jj}KԬ ‰RİƂSB@:aYs 𙰰s#5WZI_9М\ȵrF"FMq_W(N6p^mVOۥ`ǿs2.7?wۻx!9ራyқ8᫟&Z#/ZOo@n//'~' }0c%0X.|J%n TQ)ÏScJb^ڊޞ0έc9㐃R0>;t5J8]ڬ2MI#cJ('v\}gB?`Sd^0> ~^l0w^^hQk@@K&Q5?y3{{m! d7-z4uV8'K~96DZSW❚ X=ϊAZÏӜ"5̝Z?Er3L/}78o]IN#E<}_o'LtXI5˗>ۘ{WslG*,Fſgqs˼Dt~vhER\0Lfsh(eԣ'}).gaOnnUF[@~(}?Hb>+ ,KA_@bfk~5]rWjyo-r7>}U~<#$Eu1E/]9uH.޷I\+!HjM腦2u6\ZsٔX+$tW,_IpDgI6QC.|Θ* u9VT!߮ĿT;" 3J s:STc e;ޱyn0.\)\S(uYӺj$VxM ҇z.} R2wb46\ :J3HSN}b/wDpٷ/~hXXj#Y7/d !, ^Хfm[M$b >Ts0ײ߿y88Z? ν`i< cKdΐ\rkAHK0I]Q]Ͼ2? E|uY&?_'77Y,?s~krV\T)H/gWT 2o4off%zj!k?F +^v֦󷜐 g:2 l㩻ȴ?9e09tpEx #KJ1l|o{'᥇RA5i;_FʑFıP!elx=%E{&]6Đ4!LsΕs߄|F;B8y6J!(3$*@JsRkLe0B,ṀktK$ onFjxdbFxZQ " ڝRKQfMHN1e7*qMh{ˏHڇc-05S۽[ja{,I&8ɶwacCS\uzT;ym{vu+͎Z]Jzx%mBlt 7GKIc,x7{F;RR~y7\s ̮'BãI8aݎMT2;$)0'Ih:cFg;"JZWѷaiAHyUKMNA0M񗶇4pr {XmIE3Dɘmd9WL)Εt%\,"(}5d4SI0JOaI;)M~rNC%i.xs\e *D0|hTOpʓ1W!iyuB Lq>x 1I3e;Yw5G'|(9;9|.YZI\S𜥑e9mV&SvB}1T!mg._ ̹Ri|A㝔ͬf(஭@  aEIs~bu<}Ux\7x7(upe?緈Pѷ!D_ˍbB$8Sz!^l|bu.2#I Ѻ ncRBah3|T_G?~9UF`uPQ4EoܲPu_jo-|pwTA`pgYjj8.,ez;,\#% %Co(2&XcA Iސ&FM)8Bs}jq 5]'VNxFy_/'Y?_-vwf1"]r٪7\g.qM1W $@92\q& ք\Kv:xM@c>F-kz4_v?ףv>A>q`((9Z|5'y<)D9;^hjBz4 h\DC$fsQ6Dʣ9MsZ+`s."0iOڃR8^ÜH\EFsݗ>5L#M.,W9 )uɹUplHȅVK5M32̂.5`8 : 9 Ăa) ̐\Z/JfBk)Dr.<9P@Oh-)Q3 L6*=?7YQv~~VwOQElPƪVuSv)LP/8S- BQxs&Pakq9ĺv-.9,%h^3Ň&LAJvȃ(XiAÞD;d5d LS $N`ۂ<{3UZzr `r &ըT"+|71~XiY,zo~яzz.݁WKx@5rN@0T{{YyXjW?M`cnPYٛ?O?96_p{y [3gB߹4ŏd%wWpv:磀d\P QGW?&ј2ؗl{[&&k,Qٻ6$W}!#g{`/N΁o#q# I-߯zHCrFgW!KtSU52Q*_ 1-m\nA(EehcK5fLr C`SqP5DbD>zlȒ,JA.x,-qxC"xSxCY*)ܐ8:,Pl2&`0Z/q$ah-NŅhH[J߷@4 fI6$]|3S7\ޚzu5-&ex#[#6҂(({͸2A-E)Y ~ˍ7X.v0mx뫳|9)?zp9ҏgθ{D|Zf_O&`1ϋ=[5ߡ`"lW̯$ 7|m'&;e$d\F.)Ű㣜|ԬmyíNYLJ^`ng7ZmVZBhFjWk=F CLLiy4g$JQ̏ĉh-VmosZjGFcʴ`GE`zT(&֐~Γ&P_$-;6שJp؋- |ū V(fmX;fp=1. hKۈ1eNe3f"؃ꌤLJ*F]/6 Ņ mrD@r ׹IKe5 APugA+$7> kq'aGJG #"0#rlt3?@A([7/U;HRR%UeZhkÜQmaظ/|zwʑcrA0>aGx] >ŦSs9D".>]>K$Rn}a2R4 .6D-L6C1QeL+˜4F?Xnt=_ TJ g>B h^J Uf I@cAZq3\+(;ʣ\k9qx3,'8$k?S$GaFC\_Z$.edM'&a:f@ݶOgxg8ߎ;YJ=eh9YH!֐<)ɛQuՐ@X}~nΘ@ahkߜ6.w}>xwp=͟׌ /Zp }؅0]x4} Ot&|n4'&˹^+c,תsӆW˹ֽs8De ;b-?e?Ť@GЦYϫsm'f'5ê ,l;+蛑HN[1z6n7퟉4BK OnL[u3D&;HD:si2ljuIGOC[#ft #-L:dnL!tX4<9f+ 伾éD |CbI$"5$2/!Q{>99p&77o@&}ɕu[FKl?m1  j[|K㒘SrzdN[CZq0*bF1SI);8#DMRqՆ8洫&=rӎF8E@r*F o[lQ4)rb΃@ (9|2m gդ]1 bhOf`p?Gk\\.o\%P-t)Yu/3xeBՉgt,ޯmz9ߏgw<*߫||wFc!qQA< QmwD GUePT= "{F;Wh&Ktn!LqeY,d2/z'Zo=ܚW@Hi%*C)u_"#3Ood7ὶ\k:Fes ޏ6{WO1a):xgVhM0iEsyp[禎J;, :}9 ھHj>W`@ M{EA,gݼ:d&C{"K7%[5:+|3mQfss,VzXx+";< ZJ[brJzu7Ym=(Tqn(&\0xuebՍ`%DN3$cآ2Sn[﫹X[D\2QT0#s&1FA5^bB9v oAϥGDIJ͏'òh$HPahq԰jY+7ގk$ϝbA orf6a5^4s4| KEB`+VÍFP`Ds&:$"Qhl(n]k RcHbXa(^,{0bZqxo`s`J,0Da;r+-LPNXkqplv֚0 7G>|3OG?Pα>@Atvx٣MT(3B,8SĪ qX[مf)Ӥd'<3x5 w߶_L)YpmJ#]%̤v%|;/|IFZ"Ћfn!}` Rgu7J<]5 -Rs3% `\Jj<A\틱ANe)W)Z^MAwtzekMᶫl2eJ!_ˇg9ubaNgj2&'鍿&ok>xgD'\h"T cՀ[(xX=28(Y?@^A&Wy\+ZyʫV^ʵ+vB:ꐴZLro  #&Yk@W *Jcq< ^\?/o'qzP2A,00{=L܇7XRm62 СO,&b۝c)QDFӛa>ˆ|b. D;b'(DIv7%uiޓZ-Ē{=ܯ d&V{̹RH hmSd~~ _g?'tvHLY (s&=NHcs~!"AL*2ˬ8 B>jx wY tW SAsIL@.84M c$p-;a"Nh;9wH3C PuۤM#w6ɠ%Uʭ`CM&èt]6&@ٳ5$&@R>k8XI؀&!G*8Ϙ%aQ"z =Ơao7j~i6iLsd $Lyb`$U Ra Є$Ef>T+$?y)}I5&XXIݖd# w7Xq.0O<q35jSs"Z?&c+oDӾ3WZ1W9K cgcs&4khEwQϾ̿bne{()=mZ~Ud<%*msȋ2<7L3Lmik1V*Xո6VQ7##z>VퟍDEQWw׫U#U=saȮ{teXxf>̳GQC q띙(‘Bҥ`uC<]$Vt pQC %ky`1/^ LQX%\S)JrF~ 7$JT׽ ^O|mGi*"aZRqC-V1'OpN0#>+G*0mnf?y*O5xݍ&/4o7HԌtU ;R.c&p|q& z4]wo-& pn嬞4ƀa:Kzarf8C ,|8CB~E m/ Z*%(|k\1h:{f Ե?xnҕ=FkVl >G6<0%br|WKZ2[R% w (/[ [0 U,`5ۂ٢}% bbFğAğyޙYMF[?9 ЧF>z2 JY'x8O8Eן'{!gJYoA*RO[.NV#\j"c(Y`a4jK~^}#x{nRª^N`CW9eq%TsHBRNm4 E_teHu_܀[LH|6JoI‹]z L8}#zTJT^_0$1:` ycCَGH+-Yz?ވq,y煴AɅ|BvO$7vlP*u e;0¥U›@Ѿ\?m(^\!XRGs0VLsi&!Øss CȭR59( Aܶq2SfM*cdAXMP\opXtnL$ +JM f!jAy&t&L[vlee(L%뭖5I_]\xaVfsۀ9oVF3㭠‹-wHyk)U(lU%J/ivVȆQ߮Y/O⧗g`WSz"%\=R: -Qyv!>zU^4L]ypKnE(if{`:1@O tr~F3w ;tл&2U:s_4->^Xi|=%÷/2oț>?arHႏ~% Vqh[LCE?>˵'#ChAӷ'7߁)'!)`6Ө}7L ;sHc0nc1VV7ZNJ J٪A^( GftϖR𧆩l$s5Ds@-`l,N)Rr 6 DZM"O`gkLTAA PTؾ dw|V.GJ>+OJ~>1~=7O+%gv„ߠl*}G/k.Ua'wE䳏==vPCZprh$ʬ< akIxHդw#-_Picjv1xyǨ!c'ϥlkEGfUlL Y*&!k|JEpC?3SsÃ5`J ~NX==8URDj(ش")b8FFQ z\kUvː"i4*nu? oE)}#/a,~3ҙ>@zVW:d^q E猽bXik N%sv*ZR"6[wf?jl)ZwgEl\bD9Db& WYZ Ω Qq@XsG,`T- cV>W /Z$cg }5H*FJIb57PQ[sN3UZ  kԼH!f]tF̚`Sd$g#3u-=wk~>% a(wtNv"0 Hx@* =4)6)l X7sq–}`8aKXi)q%kR[nk9Pb c!P@xNŀ;m:4HO ݨw4V\oKk™"G~TC`SA׳KI":zWάcșL &!-Mr&x|RncȜ.$/.e.EZqK3K͘V$@UX! |֘ukmD<]\;,Μ n2l.l(L<1#tAFuM|TݲguIz1XxSY5Աs%wL^b&ɚr&13XR>P1N|V!}!{S~zv]ʾȈ(|k3?֤ۍ;G)c}*ZAG1lmh}#!ޟ'=JYI&zQu3ڒ?_/o1QyLj!UffT' q%p^}n)Do "˨vñ) yP4.۽zzpOcn)am1_};Ow0HkAĿ:Ӛb,f$|^M}&pݜ+ɸAI!yw\Rқ r[!hzf"sz洹_tq b"E9%:/z0sCpX$8 }}وrf}Ȩr>Uc.XGO6<t5L_y$l3%Ћ&Yq~݇rbMْ԰bznucSơfmɡPR %dPr0TI,mǕfw"HSѥ*s.5ϝ{uCsia1*R~|G$6U\;]T.ήIQi>h9V)lfe&KV4پMK2)ӿ;&%i?lfěJl%R( TοTD9JGbW_~TߟLXNDY[ uFr ﮌZ);̒`35l:!E4n?@A'{ǯ%殪sS6>ߝ-4WNV^@k(&xm2iR^W_ˎxh#E^.o݅3!c43r;(a e)W˼óPX}xEc fRzacLdC $F-\MKxA4:mAí61\g"_U)/._/FV5nmeV?鳋*:xf <&đ֒6\zZ̄D*8hʭIyN Cn-F ڲ]r ܄Y3yv{nBT:4=nYROXV쎨=Mq ~ǪvK'sE_'__1F pK:::y) {1)N|,?'ĝVG.HDGÅ5ODjPG=Ϛ/MW61GF7G`OeDf1θ2ka453s->3X܆>URb. tc4 x]޶CKK'9Up 6!{q\LQl8(ňGF|8J*D1GHQ|p!C;qkq/跬xɄ 6b]ݜzpJzBui|w^PM+dlCROfsX-e_jtrve"qaBy *„`5§Y:ᤣM:@zԹR93) V*>)“?PB+%=&NŧWUDJVe}:2lit€_EY9j~ Ns@'58kB5~aSɵʤR.c\:3{" yŽmUr&5klbtfU1{lVpl4j#9ϋx}AK6`$A73l9[J9 |जú)**AB_K ׀3/y/¡p"CGc6ÆzA#1[ΐȍ40 %Z[Hq8ь? zޗ,x|5y3o㝤դ(w0Н`1߃LI^ vo;'Oy奿MA>孥PYYx&Q"F$#WEU1rUd\U3F# bUK#SlSƪw-ؠMȍsx9ɼG.0b;EIM  3&^A~EE*\īE*\īXa1ptK&}0Hc{Wyzc| yΩPE%pۋ0GW++kQʉ(L׹)蹿>.?m\p˳f~q=W_2ba;1#tA<&U&Y#FxY?Ѣ i1٠5ޯj;Ӛ8OF)&Q>K@D9,>c׹TPp'/:#^l pcXJ ob z/zG|Ifd:+iYlX&SB\ݩx0 +ǨT#EEJ|8TFdMGhFZsr0A(-,09$82wBI"ӉAT ][n}Z΍e1[F鉿_JSoo0ǘ>~rŘOS0 "2Dw`ɇ^C2_,xw} |Wwbt6]|f|1?pŸ=0M΅FpE}=#uM bDk_4vݝ"M'̕2&hE0ȾvZExB2MU;i7 ; I4AJwlʛVZZ^%xO7q>]%J@#'7-&r^K(wS $Kx{CR 3?ۛ ֪d|GB 6*mJ5]?1f=x55ߟhiGu@aN=sO1#;~'֮hw4r#scEdP <"L00*80 bAett^v*ך0"2WyJYaWƙSJJi4&ҀcDrJ€zfmd`ЄO`ނV6a)@ FtbZK&('}% ް{qF:b/}{;3x.X`F,dVK1 5 7 9.hwG(-pqW^Q7DBڞe_jӃol*G# tU4GWOnmO+Hs0~MUA풆 8g za1?"!XGY2u qDF{P؄n:\mi4ZrL] q9בsƊ 2 VIeZS8%( bv V ă!vrZrF6]ߞMʿ/ ij ދ2u}*P|@MEs? !R)yyn0un9 ȽԹBRJY'qF#JyRI.(OjLd4O )Z 2:nf (I+~iGdE<~ Y13Gl%,)[ Ktw"XZ_~F{C҇M`]vqk7-M(E\^/z8M+!.&&"({soW:Rӏ_+6>KHpf#m2[PEN:]^:*ݻL}^%$MWHwb:g/$,mH:ߩ":ZWRsQ{Š6zql?m)F|i0RwH3*tqͦ:gܑX)$s='5JR]BH!iCA,w*⩵eύ$RsG(P8fS'eOb74 A1H x8 3%!.]<j]!MUU5eaye׼.rQTAЗDa ox `N$拾^  }JmDM &~L-z6)sĦZ8M)qa,H0*)9jR(6B⋪六G9 7Xl#%W1 Q$vlXފ Æ0fxꌓee k-V:,TJQ H/rr@VAUDBvvBhcؼBR@Hq [A6 s+N񽰱, RZ*mG JUq~`č#|E* fvreMt0S9=W^fƂyꍊՖC7mB58F<$)DŽ(HZSqQ00?һIaWj RMg^kDa1W\ؽ6濌Oðٺ0Hz#/{6Gdvcfj~_\v3:7`7wAȺ ̚ =xP k>|˶\}e[bK:IJm brHisTib.1lgHTęh*dT`\|B$$p] o$HݽNj9î" lg9d̻dR"u1ZN$$HNIeNZ$Q^ U"R2BmZJyߢY~1GlX*| 9|1O;{K($i T*B&2h(P-sJbiP J34uZsyNd<=kbF8>V| DHW"DvBAH%{bv`q{ X8NbsJ@g$ި";Ι"`5M7̰Hf1TRM !@hL 1Dpiij!\)~՘c+eeX+cR zXpԴ?WP):6qkt|s^Is+Źu=;^'1NssֺlqX.W+5c%BZJj]hLiW[9|{Βn=/ǃhsEh#l $}z(j9 [_bDN! :VIfFX+^"+(WxH}1pOpA|\Gy FwS$9%,'G m^η+G+> ,(rD aÉ<I7eUrIVra]'Bpҟc嘧F /b\"8ˀ[C]LSԯ4%d=LKH/Am62L!GgyGjN`K\T3ۺ,jA=jO?:[0ȔЗxu~~,MPb7A%vWbOHgi"sB+aN: Ō; THTP ?|Ψx v*FUM`PEo!r> c jyu˽4+;2_3>xWqv(GxeRVb̊v,ê;bcĩh( y-Z9 B~XaYEh2P 砬=t2$yȘf@{+v-zWP͌2~Q(s59TOVJb cDtA 2#I$Y,cXE|*Mj2UfսY07V\>܄: bBqUP飞ٯ~\9Kʯ}||z5#nufM;9Ϟί8F(P_|?g~Xo$ oDns5;;}MX0t`WKW>JX%NA7L2?]T+C\c2 F*V%+rQt31WAHs' շv0eo+חao\0;,2WVilHHw35 e?sz恥 ^>:˅$nXwn{ g0ïT_#?U -CP}/UtAQ@dWT9j5˶[؊tk?h+&w%ؚ Bmڧ lqUX`g [e-_` @ItC =-DOq~0c9r}X^ Ob~+wY,7* t܄LHE|}+sZvCOs}VP95ϳt |jĺў-' s{ 9x.y\~\뢷^,T?᷃x[Vc 󞱵dkӢu6-Nj:Lnb1ctJ݌BFf7Ao?GҪ2$W isײ}Dh?OY),l:RM$^gqi$6NEڦzMHۼmx}z$ [ưmeL'dV,8_Gw/zNK0V to+Ҋ˽Q<[w7էKnF1u`҆ù1E3 ;2a%)NNq ȐQe 11R!k)Y~/IFV%~rRcHF0A^9t%X&u尺]0ˈ:py`S'"q˲Sʱfh ؿlzӇe> `6\klݣJgẵi^ݾg|>g7Z:{rh ?Iۣ4?싏ITS^ॆD{ERJLC(we:Ҍ6eD!:I#Sܟna} a{EhD[fCヿ8!EJ@X2( Igy^K5MA}Q;H3ɀlLs_9agr/ΛcOҜ4iei[Z,}.~|(uٺ",(B jF)k^1囖 _ )=z_V 1J ȇ{ nM p "s7 DS#,{7[ˇtp; d* ]oFWۥK'_60dәD3$Uό4=JM6ؓI,Sdq/yyq{>ͧ\RPϕ$5cͰ ' z, XG&7#UIiRJB+,CZauI6ZIcUFTHI$9rzwEN1AQWPAfV2iJKKhW)T91!O9btEX) j".rCUE*dH8,%CN = 0ye12("E*@Z")yi("%&̈x rR8xÅL(AM-xi s)X@p1 1aiQOY RPOOMfe3L0"bAטM̟~gwy^0z|gvƬ<=)NC{ E3 y/$%\O?|8ws*+&?bRMfc~eywΔ =yWҥa:_,݌z}>B.z$ ?f.` @V`>x]K=R)JM3C-槄ĝ b.IIx")=g~)Z qdz䀢#( ttVfJQhƋ"ր᭛-DhYXa .; K-7͐rCK!/JZUs.esV}xh(*e1iu3f2Z A[g|!܆_!{6՟ZI0akx᱾#v/e͏m1]:f;{FfM9CÕI>EDy3 )d؊ sTa"}ə@Pj17E T.,HnJH )DGҢCD/ۚ)PcœQnEU,nrYnxUU&l1fN#WĩI8us~K _D _%BXIsjYEpc.ED`g5X łm]YV+q+N;vRܔN1fbSU9*v>y1+_QOfnEii}@|6]±~X Y{'[hnd_y{|1Ys\1{K|b~es{wl~x#g?s/g^3/F [^{5z#}EcyNzxsOudslI^+@ vug20Y.;M&ƿ߸8.@Js\\vcf,Oo2yΙ7yΙ7|f0so3df*!W(r+%YTISF I?Pu!!_FɔeNJ8E㔪@JեT`҉OذZ4W0\h,?FP#IOM ЈDHnJ]~WiA kPW 4 鲦Rt+{*5\`N#jPna!$ewQLA8z0AwKةմX>?<]_y4n4|$%=!^IOȯxa7q6 l*#p' Jzƈ*Y QYxN[53v8P#Jt5S[PTX,c&poꜦf0^d5SUgp{jpTS,dlՖV^tf*f1ۤf*%:j-"r+=9~~L1^\o恛l'X5DI^ +3k&UREDJH a ʉup`a]7͏Zc,5Bΐ}o"[k<1Š6[uOYn0tjQ.1Apctrf'#}rEEXS(BP׈"|s*l+`rPy2Lr(\+Gx2֐7X?!X?V#l CZ .X!>r:Jl%N6}^$ ck}0~(Vcɛ x7onb2%#a#KzQ(׽YR0%P`F)8~:'#}0C @=բ*+fE4HCȀjsdy*ɹa%FܲSEt-X TOڗBJ~;퉗XyCDb:_/-lCA_@c<eDH&d(+1;@~J7NhM9 6}EkBDaXq|Ek8V`5J) "ܨBkۥ1Ufw|0͑*ϣ`/ b3eAֶ9Rof3/e{VD'KRL:$jES.1R%EOB&^xq͈C`ULR0ۘ*Ӧ$%,rA``5#+B6g:Xq^ʊUՍkMx0Qe@c8QH EU65 `ҕ Y鼔2/$ceKVH$r97GS &1"i:>VG*"}eC8Mi5 Vq45; εwwaTsՎ親7W UŧO:awxIv3z.1aJc'ܞj5N2ʨKRHv0'ā^~\mxeݰstX\bLzP^P ~%+է_fP r2=!Z MמD(w hW$bGsvH$;$~u 'V&w7,oWQ&e~ci~buq8Rq3ɰ(/xD球#zt:؛ErZ8ٍ-^!{^=D *e1iu3f2Zʼ0UieZ0+9 dPXtoY6(yޡ"&#Y4Z0(9ܣ"ȐGX!n~p(Vx]nvag=u CV(2F X"֐|BVءE$;/% 4.մX@vp(,>Rt]^,dzspy6Rw!%|y4G #\2aȰ,ݲ7SuUfJF6SDxvRd8NRDW1@}X P ^)zCm~GumM^#+wmB1;|_j#VHa \ )/d =j{A;# M/>O1W<(E waگ~.sΧGW0_fNyzP{㔵\SqPc )Қj̍AG*W$%msØp,txx&5 ; 7BDdVɐ<%~xaff B A˜>%NMMGvj:N|\xWR}76 aD7N)(\)+tIėnP߱Q$ztsAP5K]zmϮ)bIzzxbHmO0 J"}B\ᔼO•|rګece4) qCRZMY7|}W1F9VauUVCmע'kmV/mpM𗛤 4io@򑨍E;d{mIWZ,G%9p8?Ԣ"|woH{ܒGNw0؂@ɽ>pncYK<ja/?)THΎpsQXJv(|<ͮ%mG ~ yiǏg(-w*E[|奮iIy] ؖo {\/iGΝcC\tʷUbuQt^0#Q)7FWsbRƃJÄI}pMk@ : c7Y m,^(Z^ThQֈ{:- 4Y>7xܫF>syC+7T^)8笰V mqޥ,> U\zGIK˧8Hz?oW5\mv{gϺ渷58Mm*s:}I&QmcԲDQV& .eP%hJ:漢t5!NyB9I9z: z}2}Byŏ\E4|OɜL'5wGTҢN&_.Z_nVRw(d?7PR MAf`.6 )-qD״SLNԈvfQIAӁNffsxڎv9{Z5w`[e#n5a| kU} Un&RCMPSLPkq3)T+ jE9@sR_ԓ2v$KU&0u\x9vd|8)4waB;AJn;*~gQnq;iO *41 fS] Q$sH"ڃVZ/@30Ԅ#J9&RYIJpPCT%v$X,->Zk9>f,RmU.j=UUC*E zjyș22,[";Z^(. I9v$awyf֨9#2`J߆<{cp J?qxnԾ>)rN 7v9XO1Y\80AsOvOrOOzJG8OVG!6h YШJZsR83~f6AsN/W{+<S086(zj:nQ@pA;G%(77Rjl(qP { D%J87ҠAkO)9X$IZSx& ÔW{5w]"Ҁ_.{5Q?Ȣ̢FĶRMрO~ReY$O5gC nՂy~Q9W'_\<’ 4} 0g TU.o 6`%˟x{( d%tMOupUmAm}ȷapMH 0H֔+ScҁaNOf#vnOS<6|I*MIS3matghm$o +ޜp -.36ߠ ӝ|Z"[:5D8p [#\Ub0>bZԕtrQ1 RM͌7ɓۢƫk8axyح_#i%ggxu  /2&NF R"ZpR2EIҕΡ|;X9x R:pt)qY:Z,ICV|D94j8<&nL'xij9u5eO)+^R ں硧~*'0W`k|tgRBz&\P߼Ci28Gm,B$* x]K7f@Uҙ6g Ržgkl~ L7XrW>6:[/ A [np:cek.󝩔i/BrZVE(pGX1z&KYc)I"X 떊Bii4bs7-a/+mdraLqZ]zߑp|I,zdX!.TQQT.kUJ= ,20:u12D.DJDpΞ,k+ å+XUpE^~Nyfy}+)KZkzbΦ$[T$+{H[U-φ`egy<3U&\3=k)@wӝ|-AԈjvr"ZG;r:JA5'_W;U;Tl^cz h8z3UT"UVSKAׂuao Fŋp}P҂|YWhqӗYo]FslϤz=7w ԕ Rk^12BЈRVv,T6Nǚ:Nb&1 f#rtiʫMZG uuZULq\eF՚hHҀA((~ѻ SZV((OʒSU9`wy__O>r7jQK %RFSb! D9#l4shԬ q9YEB.7p}SuO7_&Dzﱜu=C姵`󫛿 ?c=@uWu,-BU4X'@@#<|Vcfvw Q8$Y[Ɋ1wo!K '? ދez/[U\S[r8 +h#(-zeO7f[?kA-qt/GH/uѬ`lRNvF;^,Z!G=(|YfZ1 QLEJQǐ'ȷq;£'û(D 8=ʴ}#{]x_Be"g4xU*da"UXlѩ] cho8<[ST\E=]( 5o]ݨ 0Kv Y nضC@֟ GDP7 T f]lMw]OHw(B:i&#a@R_hK(q%@7eT@-VSSbVV֢ tKj5(wY~=FL7;:7ePTU^++*LG1Mx:0 m:}˯8؀γȳ7,罅ۤ|j]D0*vpx'$B q !7!SkJX-_xx0h۳.8{)z=\CQ0UFNu%|wx%xaNj*DLP{ۄ0м?AGMUKJĮG!ܽwtšS("H)XXJIj]iy ;CL}?_w]SL4h؅_2LnBD.YI%6Rb(%6RbӨ4_po˘\F<2" qlRآ6״S?w( q|5w4tڒN8onQABtA|}7dF0̈L.Fm.FmTdr3ǸGADR8_r-) x<kXK-p nttFtYg#gΚ"W :oWs~o/g_?{ӬLw$6;?|~۔x:DNqRӠӑDKQ; RSI=Ni> )N@CypL%HaL`4T«h)w)@dC'`M9]xGĝ ϳuq&#5ٰiD@iKix{N#j:Jh_eLX< G Pn)-O"mT;v*R$'4VQUTc,Cp0:X'53m5mH@B%^qQ$S%b2?L,ϪFTtuM4 -<@ )FVh.<9e IA,RB#Fs# 1ADUUЍ^uͿ [+JЉlXT'[JBQ.ӢV =>7cs%ϵyvh4I5 "gdz/0+Ph)Ytz#{&%|fbހ RѺSLfp]`61^ fW*..W[eem/cHɲ7q)$ΠoVtɛek>Kdzg*dF!VvJXʵ@䲼ddN_m"]|VXDmӻ%(v˿qE,"qws2⾵+o'ЌIRm%͕ E|./3j/5ZJotFg@%l!LD7J ]R.hXe re!ŵ\rѭSZqi$8-MV`(R:/,_XRn RQ3F\JjE*a1 A!ȐZEXxPcrqZKHC҈\F<y\<4G=ӚC* #Zi<>z17\AƎ|IUP(PGC9>Uk& ő8w<(XnBD)(]&3z2hWRc)?Jkƃ*2%%G.8[}7^G*荕> 1Zp@ )>Wē9 ˝ &S f&䧒Ǭ1+!-,q*%?C< S{e]=d=x߽HBeɁ|ό fG7Ȏ1PCy= &ó$S[tȩʽRPn[F9nRl$\1*A5A"RLث$?es xKw MfaܓDqxqB㘱( ΰwYir$kaI]Ǔ(,MVVqit5Tr%cZhLSy4<ƙ!Ðk,@mNz ٙa%O )&hBZ :)2D,7H'CPVDf*mʧAm3ho~X;G4#01n"uQLgz CfwbJ)g +,itu6k9-;x4"_)b\lNaQ2~l:1j֠,>>=\yܒ9[[\11m<&`ghhW5X<>cؖ?Yڰ ,~w>-v.GRBܵ¥,xwFIp}wg=sKbSm:W.'xg{ 5zkCL$o>|NQ{~.;B7!zMpDEpk0ycQגÓo֒f|ńs}h:6J(6]a,cb}hvt4gIpvt= h ifǂL;R h;#V\^cMg#ظtOUֺeYYθBں(<ے1Ny\+<6~8O\-m2 rQ=IbUU*$ʥV]W2BsT^=RGeDz8*^v#B40;(TSG+. J-L=Q_9aBrӽ3 kp K'_;$"%㹽_QY^TXp\,gFr"o^R3w "/)5S)8{WN%˕T(L*aRS)"HwQ%1!0vj~l5Z㓪A@#4@C&KP)q[ 6}<ߜMrF0YdkM q.v wOi'}d(w~@-Se$&T.ŏή|OFE=20wYAp+f\^rF !)CzD[rmEO@mkĝ^eΥILzi׈'Qam'y.P fB6ٓ5L#ϭcϢ-K1=YkO#ioZǷPj-ךߑH[vP T>ǛI MbJ7'nB5GT=ϿF {D`)Fnk)S@1ߵR01N53+ři kQ5i6ybs,bJ_fUы6iB=6`98N (S`6E&%c+Wd)TȔܕg'?mWbNQY!Nc5/:r CJ= %V=*⪩#^M[uQɯ&?^pb 1L)ZMv#Mz瞾%"֚{ &޻<LfӇ#gV♮髩Y<^]1Ǒ^ }CJ8׫0$t8p$(ui^|Y䨀辣s;{]URb%fpXƅ7N;8fhM;a FWn^t?|y Z4#ys9|* {qTh|P q): =>nzUc3պBs=׊aP &@$!i%)5HJLG}[5Cg&QIA38M-{|S8RJ6Nr<%++-MS6[58e].R0-2%stʓ^K&$T%jkM0&Va`7PIF =ޙ?S V2c^Fx:Ak6(@uM[ew&k˒r]j2$aDy J)?S vV(!0ccFHRY")d05 9 'dL\[Bޫ2ʋO3 (d]Dfk/'%B)A v\ab$J@!zoS8(װ mQ@ oL` NbJlw›(P+ TfqlHf[rF*RxI 䜢*/.Q#N ke5B l1&}r8am &{RЄ" )*hE8ejAaY k XgS`/78VR ̙CˠS)N q&956rDz`gSDYU;@_J.}O. |`45ȎTG{O"USLYLH) $(֩[ F+*Ji̩)1<M:e"McbVhșXZaY~Y;}R j'߷:a⅔Fu֗YYUyH}kD%rRYTDmms`Lo R A w;E#:@Ύ;fc Q1ECTv`UD[JB75vkowkopS_:or36cwGfÃc fӦF~?}}ՓqTrhwʚ/z2zju5YptfRq+<2A@#|f6>4f!"v`amr,AЗ:!SZ.͋pLoRԽT[ЙG^0W0L2ɊwGL8vƤ NSZq A]Dz 5Z:fA1et¨:d;<|v\?*odVl|abD.b ̂Rl.yۺ ĦEȠDӄC3d_9J3A@Ig;z4zH\bE1 (A$P T%@*b@&Z2uCOBڭ 3Jxi҃{UHOigJb1C']YpAxђ !(Ybs\\B9b0!H$%"R,e@ݔsp _ ::5JgJ)¼ʓ(EԉRk2xvjD |*-F­?LGE1ݨZ]nlMozU"Ssm_1~ܬ6z]+bQ}r0yO\Ǫt1o*^5۝ӿxӪmݙ[IOxW, V.VGB^3"j;j7)+66ڭG{ci7⎖͇ru)[+wZmҨ֊j炓ݩ&A17ZhOQчˣlwzftFSd2/G( ;ضB!$[ifTKUWo;ǽu\\=-Wdvx7/Ղ}֏RHڝ`~ x!N3/z^<U/t;zT Mr5b1>w?AjqQѮW.a.@ #~#:=#ۭnl?XX튅ǫC_#ѷ5,Z\ W\YׯZ=]ԓ~$=s}qG]2J ֵHfߘŖ$a Ferke3BSۨ6 Iϓ>6yZq@,gfOhp; 2^6k1fB2z8J=Gƙa0Vg ӗ9`:Q ƸCnQ-C!bwN(k/.`X=l_$HqKh+Ẻh}QPL(׮mCЩ3=(@Ppx툺[DA`)mc}mrWIDXvvmFFcM6xT.#Y_Hz2_Ha̷!fi>v#T\CМb:#B >$]Y5Gb- sZ]ici6#4<# 3r>N5$KH$2CyAHr|6#<"wAL$)gc"D1MqB"&ϩ;DDVD`.▇ ;IDZ_;`;XMoC""sG$=nK̳˵`P$!-$&俄D~""1t!P%\W,scV?'KtH 7w [YI\e');y!I10EPR.'%"L2 )3PЌ$$e6|&;ΌwZ}N(uS*\LJTG֑!p/9A {:8~@K}A^JJM] d;=M؜lDasXn1r9k97–3()YLtΒeْ"p%%9J8U/0B% P`N~*!jZCo!t;ڴR<2ZKs˶>$䕋L Q9[ݩ9ӯ2`zYݼN+mHT6&ξJwMb X M nЗ=>}&nUЙ@ameNu.N.Vvݰ8- |磷>W;gXLfӪtװ(N5ɝ ]C(wrbuu"51N<ǮV eedƔװye] \|^o7pȕ~fR #HOQpƟ*Ыu;wTy 5,w,_ "{g1N̡Ⱦ&T.6(v sQN 7.(,_M>aJjeS5(7.p#y& zpZ?CZ 65qN,>τI<Ղ{Ԟ}8qT*d$"3ݕsՔa35/uf[P=>~Ыԟ|̦UCsy^\ '3s%s]r5R]FVw4\ ûz8R^m9:Py)ڲ?IUipr<8AcXxLZjGd,J$ 7lAyKF ȉVv? eDwHb'FPD~.@`dž䧁>:DI)NX~J' "/H;N)/5ɔQd30=Z`9TW l\&U}SZW..2%|v"q}G_&Z[j7EI8СvcF&Vq}G]$Ѵ{ KCB^roM=j7W lDm[,^b5͇ru)NCHٜ3ϕQD*3(q;8Or}?+N-RQԳp| F~«0ZvO?'g?U X"zz~s%( sr~ו`<?0#'+qz~'+p ~O'(qGg?J-MuN= { fh ?f1c7gh"VgC%Й ޥl@Cʍ ?,gdKl"Vdϋ$OB7)fwjOqJmB*k=7YId s9TW\h2^ ,6_1jJD8@F bU3SYq+pk2pX*0*- mGcfc{H3ٷwˇXjT<)Cد8׎?h_]?oxǀ! |So~f'e^V̖5V/VHd?F31y^zS£'/p NSyDJӲ(@ZA1AN)F M~$́hh[%"mHw7K1[>_YߤwUrŷyMfy_hE<$$Ka"̳9%d0y±.`R"KY"A889A]လFAd$M`6è&O *cIcPf) 4WeOOH $%R E0t^Oz\?8ҷPE1) q9D"zT՗d~u?I֯@\ٌp~)䭼?O$0eeZRhZ'U穓˱sL 2B&,`B)́R`T$ ˕=p%rpJBR NQ""IRQHI&҆y , 9BBP ILrܐ,rD 0)˔'EӇ=!յarHi(i.JZ2&DN^*v pvnny:Δ,Ɵ ~*-}Tjh4%wEe6^XRԟh,:m9VJL bqj?NiO1w+:Y,&wNYkj&MTå-oV 1it0,a@)B:ZG7<_<-3'!c=')1bTs "ZjKy{R (dLNxA 9}^MrpGaQr ɡ1s(r|¼HҐ#XUM-1Z@B!L9Ω Le"(L;;K\҂pYJH9i%Q9SiN 1$ctV9dj6ihH]-jU[)0;?PDzN]?=V[ >o?|%1MA֞>@?X,It}?}z㑘1`IFmlXSwBYRITH6Ҡzg2]⧵Yze83QFI) !@aLGwkp{cJ 1#FQ4UGK(>F(Ax)HiMW2]$Kj3u@]4{u?%o; ;J%V^ZaDPҊJ$DzAЊiT$Rw"ʹt'j{݉j"0Mq?^SVkvA?J#{^Ъr۫Քj/\VӮ{E= ^k$ժ@Șnn[V@ŐySZ]Xz&~tM$aKX|x!ƣQ$ȁBz;{ph(vZyn0ކ/eCojsdM`Y>ڙW=Ѝg0^It Fqs j\P=`L꣠#R`$('z F%fz\ωWp=oou sn]hj(eQյh.ı7QIp1wY¹Ī4KBܺd;kJYxż)l){T( ˴d#aT:ansxTt}q#%aF!aXcuj@cMa̾3dXXE edmUɥia> Ra8u Z z>2[ BUXDr!&yxQꂴsś91zuWhjWVe) g Rd`dD<Hj\I5X:3(;SU)[%pLhSB;3Hf`#HmKEZf@j-(-)}Yp8X/ r,"nߊI#zzBR͔VGͪږhwiu%]zo[nڲG6fJcWĂ'x]L|pYeM˺ͨH/eeUFZT[(bBqX=qs5Q8hơ XH\JwYlPE7A ŗ+NKmZfՈ~Av::pD4R؎\bоO{T}U8t?T$ō)ÚW:ndm0'?HaZ( _#«߼ ~L:p+z“1q=] W RYGgY,@rI:`ZzM(c ]5&4WV0bBP@ZɨidJ*6*Zf-5mGeCdkF&mo!Omo?Td .+58 I俢 J^$ ;btqG|ek&Ftp $dIL.S4d;[lu\.-s[w?b8 \DX}io=sdY!j-æ'IR6`$$=2T J#2 Ot+3la 4?$&495FM;(FgO'<6kn3 8e& w4}(ЊTse2KxFYb P9EWx*3x՚"a;E??WR;qyeɼ"u}GC@mޖSإD-#Uv)ңVzo' o\+` DlV|SX N89C$ Hϻ"zlS-BQx$&PMiW 4[A*''eC}Y1ۅ-}7R6nZhkW@HNPG+^!7-!nWM1>o8@3Ez8C#_>2D@86Cz?ۣI SpH*BD۱ #Лhp[~p>Rދ7\`Pm]=TIsʛ%\K;8]ϰ(}Pzpþ!h8 ^0Q-ZG=0I ЬSqM eOnQ*–sqB'`Qgn]_ps;-U"4?_OSb' 1T~BW!T~BWPzt6K2yRÉKsGyJR)3r)*789zǖ9F ;wBlލB=A6˷*e|WKMGӬdmЦfl(13*]CpwU<+j'>s>4ST\iD66 KG9exA T\IGsu_TK|K?h6Ӑ90jp^y{~t#q܈F\݈V;/VIJUƬ?`)0ʴA˕s0 %VHj/u_~VKI6YZy+Ŝm\*|ÒϽT)Qĩ(L)NuFYň< 7,UMqbQj)J;ѩFC lK4A3TLdXɪ'@Wmm?/,f>//y_͋9G;>&){SfNe`6m)W.&c Yiύy-9Xׯ%Dvjg)̘K0%@z|qSLjcQvaa\1='zcp4RR;aS AU6~pꪺpG7z8Q WV_-D~ KǏϟϗDž&|*`I֣1ʳb/_S,-sm狟ܭ->,-0%"-hR]KlLMp0VU])O,S$iM5@U"V%#6& cҳK%SCz+Cy@0!]d5zzl̑ ;X.C6=ɱd:0Jm!6pݧUQN{; ~BG0vr|.!tQ5W{,U?4/`oG(1"w ' O:ݤι,}7ɻ eB. ;;*ē- THD ;*πwɻ< [&2r,p]F׋0U8@oQ9֢+:ԯZDgג={jHg/a 'kҼ#C0&?o+$wOqn6 횃b^+-jW"|N4 g`"SM@!6y*m6 $ AW||Jrbg@h8VRlVrr<0ĐP &i~߾|E:S)^jp$)G՜c⡹bÈڧ&id?#U̿}%ߖT|vw3s%)b|1~å-\uw'b_ ${'=K\҄'I s*RGd;kG(%{K6T1rtV͒y:hU{c0z9ã<)]­GrHtD'9Rg ݨXۼjH}[|*E(xǹCLHiɰţBɱPVXH;,iZ LKlXč{|*J`0XR4!đێ 9(Gk#0D.b1@bG&Bs>!Z~˗)㇟?T bumV̓az㯞 gejvM}ԏ5=UmdqJ]9MAdF'$YŭDL,T0To3{X<ƒal\=w Tq<%pVD\X{bd>U^Rv&2 e%"GIߓϋ}Cʹ=R.r2k%؄8M*wJQDNݵ lҧxhՎWG/ҼCgry*x9a~.7i*V^Ï=kt3UN̲+[<7Y=^\^Q 92*{Wb*{!TVR熤UǼoV޻WqDg+Q oZ;l}B*2*!.&G%YzY-ܿԖHطPU6Bܪnn-BSݛ[ PsգP*Mӌa `%QDaq˜*rƑfT2U}:tja+2]%>uW?m>- '*8(k!0j\=6NAN/Xn!еhpf{wWаgYa!JW,**-H]; hlsOTeH(8TWW7T9.opV6nXs RV&hObN5]P1<һhUVtTS!a(G[cA3Lv]|pw"/:E9SJIPdNic Y7=9 0Ќ'_(jj &7f~liL I[~zcًm{.}vf)TVeRbO<\<}ehgsU/`s(jҐo\E+G-[7F1XN6nAuiV=CuCCqSl:[Xnr~ ֭*!Smg r[@V|*ZEG?^nBcnU1QmlcFB${nOZ:4W:%|z4߂DJ8Z_B"GL&ݑ%ʉ$pҽѻ< h6p; es)!sU) {nf϶|Mw,I_!Ҁ|(V.oAO , MO _I8{N] WkzJK`,' 9݉)X{&мdq΃K>#lP>K6d+G]$73Q*ԟg>"B0GyEOnWSzv14*TS,z4.o>hNޝb!hꢌzpj1hƛwX8!]ԶDmp܎8.Fc#{+_PKj1)^BSF| aE@_k0{ IvO;F aEJnm(nqQj ִqZBax#G(†{Sڝ<-9"IOݤ$ Kw=!!6j%?כC6 lj׽R{?βV)38Z ;޼̳+hR6]5m_6IH[#HGhLXa9vCE\Ǖ'\ս%T r0̀`?wt%OHʓD"neXI4T`'#.26[08lt0m%'!\i+ħw1jcQ""J'L"IZ$M" r x:R?؏<0Znnyv6"MQ  Tx VVJ5(ln _+jUqC4TNwln۱ E1g<*;x{ nE1 {xMrՊ B5];*3p |,5/_8x)LDh,ڊڦi&NWV?Y])#Z5K'vWfPִqBXsb <.SOIb9ڟشo_m<~~-LXoG"(AZH.-ط)ģ(SoN"M2l^ct9>OۍrYDT֜r 1Xg<|?@3pG_3D3=::k.^ԏsmD@BنăJtλ_k,Q~LxIT#Jx4]'v}O{b9rZ0֠I4JR%M"Ge'j@1$UmR£ʊS%tGJ<b+Qu=$CKY+4HE"Ԟ݅F \G*AYf1cFD`blmlCqD16aNXG*P,bۊj09lE#k.kA}!3chD2\wN(0T+Uiيj 38`"p,EC H41hc '#(#"b*@tVTH+!P )'"1XJbpT'HYD$.2*XKNK'O-$?`ӎY #XI %1s:2L^W|%r9g30Iӳ,u>95_6j ̧ >̰o>wǣ >:NcR(ݲ|`&/~;bQ}}EO>cbP8jxr7[!0mDB_gg0q3/'7700MGsz'uɿO_8?kDx:ҭƄb:5(}5.g#Tɕ |uu讪cn8Ppo,@IG,wRx灼C=0%2߹*ey#沣ڋ%i50T T\S%t*& }/{d]DmQa&0&ȶ<Yݚ LD]gfDQO%;,9nw{ƵM1q%a*(h0W'rת!zxҟқZ0D)>e@[[ D1.Q1ݼڄfi#+FLZ6v#jÒpãlv4&x@S| 8BwȻ_%>X ffi1'=$y rwvd|,yeM,?E_Qwۦu1I&L,n\"i֔ F@{jsƜU#F_?w^_yӗY7^nγ:~~ߋ?~{ߞz\>K/^&//zw/z'qg7Q;}=J_ʟ3n롵^o>_Gw//EI/]f^=XnVw)EzaӘF*D'lzqUŭsp4e1dFCX#%+F(şY^nM)hZMB $X~3l< $c9BpDȍF 㥶d˹JoށA,wۦwv̾rh½Buɡ2 Q'_&uG3UrL;aV)Gv ^ΰ}m4'nlhzRh1O=)=I!(Uh[*ȤHDyI8Zx!1]P>*H<]ڹQLFY46\VBnEYlDAF#8L ?*^[u:QeSƏLS+\/hGcӽ.ɱv\I5Pn,Yrn` !u>Q ,e;ZjbyX;XmB60 ϏC BpDZʆ{`㾼nĆn7hV:9 Hp?3A_I noUG -DR`.ÞBvN7b\ا#P5;&Mw|xg柞vwwB쇛1"++.r_**8=߅KV<]Jm fp.cP ]sXm ՕBFp޸"RS\W+r8+U4ܛv| CgUx3[R6^0V0v^0ޜn[c1c9qx-R{KӢx WVZ΁ž+w>p3j/ƚ}Q,TTmAqqBQg8s+ƹ Wʪe _H{:ZTm}=X3iS*k a^ BpIL~b02{Pל_~Ng= Ɯ*x3vT\׈s=A8N1]U`97>IGRuZ7bG6t JkHzTSz9UYyc[ё*r&y4* 0{jZ)`92+) j0rǧA<1 ]k(~l53rN[pBѵ/k8/:_?KR%NV1ƶw_QM/9$F BVi!7ԁ^;qFPi@.ŐKuc]r>IlQ OX̵nڱ7) DQRxB UUWvP(=kņ2 ՘oG\K]&tZcα]zWw 0 4)r \v_x4={ @pd~LlѻTo~f 3bxt)&CXL `0Lp27~O$_@jUxzD3 UR5^$k^8bRп@ӕ^JN<׭oO Rξ>=" 4I`3({ >#@1M0I#c( T\2bOXiHO;+t )x 7`e'%_tջŬ]e{&}ͿKͿFqoȇra5V\:jM9͋~ﮖꅊMʩ>x dwY W}&RDD"4L I IrMЀ$ r, w +x_GpAЕANb.W*~Cһ'PqHbE KAi'd1( 5$&H Y % Fhǹ!=NLd ;lm֪ @Lb 7XD0a&AY"JRX6A JlO6')MX(PqDI!SI2 HҀ 97J4S:Dq @}s LRiXBPH,S*T* 1\`hPE`Ҁ\qpØ@h 8ܧޞGS!յ.ŋ,QL> pMA>MY:Iu^wt'!>XlX]?# ˯?{-sj!Z~!_LЁԐT*NC)5bJ+Ma,;]Y)|k<{tۘ7]»sPXR })D1qNѐcNle((T%RpH΄JĠ`=p 5 gCACHj,75U18f;㌬!A}hҿ (oS3neh=Ɯ.sXtHR1Jf5=xͨO?\16q澲E`r]i1Ȧ15Ee@d!ҞAu@C'Zb)#M(" -S#Dc# ,PZ#X*XYbd"$b:Y-{]?ӑ:3nR,Fl5r8@(TYZr6qנyTl(8xL]>wNTfzw W(+vu}җ)'#-h~kT }62 l( oF^bqNتMvVLArk͢ɓɶmj8y6+VefH>B   T NHHH  qcDSYĒA0T$eDVD:jpeqК쉆HNwGC<%3ƕT)v(YRiAa # arEYZR^pEQޝkQB쀰\M[E{&;*ɖߤ˖|O˜?}gǦw{kzEhx{ɲf\ͬۏ7ZkAwd ]㉙?7f+kؚv757Q`>^U;|&=ɪbX%+7.d*׹v㹭3jNwngt\צڎڭ y"ZJr%*m͕Rzwvn]ȯD Mn%XH3~"Əvұ2H+J="]߱FbeHk{&swXU hYjp_FOH 4%k"ȜZk5fxws @i'eImO* M~7_|ae4>3{0 d9M,{J ^FmzT;ޤWȗ)ė{0'aX90ޙ{,/qU>њœ b rzɱϺ^ n*Yb}` +S.98Ux.OV˪5ug.EG#B?}eXHw{eȃd\y7Ͱ`J\us POIC[G 9Kd {w=nHnN tvn0~E&F۽;"ۏݶlK6e10IeTXU,VUTUMgbzpTԽkyI|5>oxSʠd{_(^ad!XgU yF)Nbh&Vj4H KrEXδ(%GN,@N!c,3@dixn ܋AtRC1D! I06G03bIhtkO90g+Ԛ=iJEOW7|sG]Eku:qI~ubx_+xQWNmO#Y0` s}E&))hwTp{NHBNxyΝ=+ڸ'e*sϹgQ"e'W0W8GIGP C27]rchTDn鏹zdl曛]zW?rm%FI`Fnɴh=!g1Bbg̙޺m]ЊNa/?y$cZ kd1UIf(m݆(RcJ"2n:ޕs{AnF uO vUȡ( ,jg;j^D ,łc#ʘMN!29ĻM(cK5*ԡ:$h2sn,2.H02B!UH8RPaӚX5l(C69a{,0a-ڱ~3ޚEmD @І^LkøQ\ 5vEdYYU33u3V+BIU:/::fL;3gF#aj2nE8+ bb'S؂.l |JȟbԹV%<Lx.EH{>8}>-Z' M{#Gxr%-f QxSbx2" Whd" 1tۻX"ǪK`{A }7 1 9(ϟܭNCFO$\5F;Z5Fz1b9N)o؎=w;h~xfs% e(O'z2 n!D'y=54tќ 9|)e78E1 BI|@GB1dhxHfU6.;ȀwBTrqI4T̽`HVn̵ӭ MOnUմhsF怄|P[3Ҽ=~".'\I|R~.^Dy=6F0 gBk1ȉ(@V,rs:1Y6ր\㚓Vat43l63gzpd6'-ŭf<{5D$"GG A3<1"JD񐈰 vAPݫ :DL)ᗓ'LK<Q`zm-uu62I] bxdheޗ;F) GGwE/F2Wo}$%˙΄X71h@ũ9{9Q{@U^Mr=\b+#s6ZP`"5& `zF d"ShT\%I4* G/H*B k3Pfp_s9t{މ˧s􉲎RD|#F>=9N+]/W{O G a| o_LEj|d9Zme0ʙ6n4˫ٽ.1kΊx9t"9s~Ӈ>ۉt` ub_\/?cIy1˷CIEmյcMǼ/uI=v]Lg.fE[scWYϫLt,EHdV.w$!?Tu+\t  iO;!>pQp򝵖ԛ =LZo,<)5%zi-IJŊ-ZUEbBkJ:?)I(6? *<|x(P8ͧ y:՜Jr'|d%Hup=; 8s?smN20b3ZPizjP |Lc&}r,^3څ""S aڭt- vK}I89OxqOօ"$S úoznNm[Bi%F[Wxn]H.!2|`_9m- v1# (v }ݺ\DO) k*~a5Z_@RLXmH,:8vF"#1"uOG2\ o՝lH Xons)~,S'[;l+?3-ԡxknKG}xT'nˎ= QbEc38t0&!%:m#ʣR"2uTӡ8Q4vwǖA4uNȘHnXξU)† dp4nc er=s#[LP o`Y-f_M7F;GԶF=W} M"s\OW3OWoDDPqG;"XVC%ې-9g3w"Rx&l%bL?Vi~+NfWeur40$XpEh(,`#DtPe^,DZj˳;3`qL!8ӌ1ԚÜRCkeN79zyCNF_ox1ab =ZFKvz6?vRjȄJR?U<֊Lk<>$JʘW$DM)X&qۋpM'@8^_bXOu#tN5#zJX~%ͧxKNQ̶x wI|^ӲH;8цݬVg}}F@ȻFAm7|ۗ;}y+Q,_u0HWp\AK(> hr-_CCjOdr(rf wΖKH5#Έ'`ZD?.E`^RtC;n7@]/1|cfhSȮK,KV-Cq9{ @ ;gJ\C XfJy߉ gÊF D!GCH%H|mW7{Q >.'^~w yA@'>z30%?@u|4/g/"R/#xflHnC\dc:&Rh+c0e={jǷW_,Xƒxwfw#[qp2 KQ>=vt.+:8 H-I6C/7yy0V$I(DWM'QA`hk7~N07)E"%λe~_: ":9r]htr֓AnJ&cd% 9R5dD0qOF7?x&M(~ W8j%(WYVd6ψp(RcJ"2nldY/A̗+7fQZ<&ow~ j7[ՀX~^_CWvoӯ꺸"Ap esV f!ÊtQ/a~\"Ze OZ `.h>}iY=PNhİJo~o=zPEtT١j^}VݩOT4̾_-YFzK?+V˫-=@ $`st0/xkn 1.wɧ[ 4}vJjhgFnf-ba'r$jILM3Ɖ5hKx }罨pH#y:VS]YE ϡ=USGg(i]NA+Q;{cίܑsp+|tFv6M"L?l]E$6wׯZ( Xb8XGGwE;+8zצVvR!Gby/Iwmmzl=,b'X:m_6j%栗!E=j(10bYtU]{q,nޜV2L?4q}yu<_ޜF9jQ'S'-fLv;3 z`2M}BV s2ύ;ZD}\^ӟ!o?T'_Σל{欔-LIy5gn8䅳h+]Zj0tJNw4nOG֎Un [YO)t!j̲s9. =ת.an.B.; 鲹v ]h+3JDmw4V֪m τ.΢mxlޥnvթHh?*tp)zK/WD 8\m{ٓٓyP=RbrceyQfki)k "lj{p$LB뗯r5,(UAZ%UvMfd0'2X{#aqDxg͈R/% 1Zؘ59+PCC䴣-]j;(ll|TSi:nmE;GK12sфR5Pҵ:= @Iz-uW0gɔ0ת\]̒zh\LQuT4bh2qMZD `C1R>ynFu q_~DU1m/Hs0m qT<RCm%<0EX9<؅a@C|QFXA}]=E>(w?e;4g4K)7$[K Mڛ3ɡv cccID#!1>ͫwEX\3|\]0|9aL1P=]9փZ?b0k/%. R=21poCҦ\&܏w ֻr܀>%GԷNkGݠ~p| ^H`KY]+&um{it9as^QLgiT UUVSWY\aJc9q4|z +u~*m/s؄5k lQb(^٭“zI"ZxjvF9n|Ё5:o?:Y2]JJX+n5^EAt"bȖ⍔.ԟ;:mğSLŗk)!b 8k)z=ݫ})%l(H6z7۷|zAhh Ϳw缟* DMFVKY^ pm+z|ɕ^tUlT4 INP\3$RCA\1zB]D)Ĝ3r0H?m m*^SKUR(H R =pFbC-[Ds6H#kkk ٶS0p%z#τͧ\oqv=EF8n ki(];x3t~퀧t6I6eo~~R{j6W8 RFTi':zAn'OOSZy=Sιzr1Pt\\yta&ʨߙtL91'Fmqݠ>Ό]L g0jY Ζ.~q2}z\O=LȖ|J ShKM[Dcz+vg[6Yz߾5eғ\ee^b\i2VZTzfĠ\Dm99ν8pjP(fqzixH  *w8;)14FD@u%x3hr B6N9i̦W@vzN L)OQb[dR%g!F#Q2%q$AEe1{x$!vupZE6iE'Y'4[%~!1 A_LKLLAe ?>p+2lh)u 8UQ$ȕp[X' c`w.2:" 4k#:XY` hʅj >?*'[!'6Zy).::k( q|oʂ w!HdVg錞us@!KXx&h&TYC_,߆ffP$'A]';!jo?so6熭gp sP"1 PޔsP\8)AawAK?+Z]5KpT_qMrFD(ghtzGUhx he j3 0v Scʑ+sk#&5 suD|!JõDЈ*&Kޚ(V1(4MA gqy!*boJuZIPUMSkl]D 4_'c!5AA>OѪbL`<>RihL<"' fXûvP#8ᔮ2.d8=F NdiChٴ|AtB%#nS ) ˟)[;)e(aωc-\;#+)δYǏ;#do\c(].R3LVIS Ϛ1kZ`irr1qe ZCQK/9P,.Q P5h|*xeƧf44$`VBB.:"h)L!c-c6б֐/fMBdp\qbOc=32$UYg vt(f5OEՌ{f}PqQhDN!x1h1\ tY mUJ<8Lj (Âh;e} ,)8Bw5G j-\ih/*EEeERʤC 8H&-f \r'Qr.7d<'Q(BF:IXj1km֚jڇēp*pANN!A[{(S=L"$)(280cfcjFuQv&{ ״,J=: J}[nR0aNi @04XAL z=M}+ "[?v+_2j2nbX."뛷߾)/7 ̯jrnjtՆTw?w?3C!Ƅ^Q}N2Xz1yښ4LvHTFYr|duO'WZg͗l@@ڀrP* ummlya))eEv|Mۮm~v a*ԏ:XMg]EN3j\0}!4T9oAʨ`+g&! iߴP[ @4$c(9:'R:8TlC!9.=%kv`8p7,r= ˡQ68p @LN8Lr⡲^@6^}jJݒ=7o6C r!FpEK}%Z{_Ϯ>Y'2 PIPq?9'TpHBՆw,רffE kVpͨ=n>*,5/~*FGqc1}8E;tJK:G}&2w$UB4x@㓰|]5XQGJÌyre֥7>o7mំalzR<]GjC-;aqYoHPm@ߚwZC;(D~¶%fZ8(*zxPݲd:7jղYg?ߨA9XuMFȡ'M:a"aқ"M$H#.W[M-.OEB>Sjݻ Zh3v< <`lZcچ$ٳ>afC< ^e˥L\.lŽy@hr]ub]Y \K75exG%⩖G)㇞T?]|vfjeFw&im"j!g c]C$=LJ2Ԯ*w/3 ?e {ф{v9D9gS_K$=z}3Kzգ4Pd@΋ buQ! ٣}Q>Ꝅ9xu!3Q Oo.:Fv6olu·4Ԑ>\I.2"MIQ).{9:و1׃W/ihׁȚi2#A2sB!rՠp Seho.XkxwsW)9IisTb&薣LmQ`9xAO,~d䉕V=h0{p5qP^Cth5vk[p[]LugZ~(pbzfw^ȓpR_b]ZLY[/-]vKF>zM[YJIٳft3m<}-<>)j9nIx49p$spHq):]B 'l'Ch8AD$0A(kDJ'ZȜpux%}V4vk5$vivȐB(6.luv.q *3*8%&eP7C| Am*~mEf\~1zP 5):;M{;2䞔2h\M!L5Pb{5">pQEڠdLtx-<8)$\EC'4% NNr%.Bh&'RbPn6!iq,5]RpKSN4jrI1a?+,)&\tS&K NR|8$}8ωӢκ۹;kdX\+I5RwS{{=TBz Og@lGw/ ڻP|ӡ"g\MfDh%|rqY6fJqQ 1,%uEÝeGf97Wcey~VX4WpHo1bVR w+i %@Y2\6F(7多?Cbc 6NqF}Ac"j,8!W+-/Pf % =MVL3];JNM`_[:*xtxA++8\BRB.R>Ig/@Qj-+i e+R%K$JG%3:++Lu(&}\X)oϣʀ| g?ЩŢLkA~p x3/ݭ>- s9 =?]|CSASIJtʊ ҉e!5LJ S`$E00 l*l˲( NTnV4jY}Gw=6Rw]9Bjh էZnT+`ivHFu@ɾ _u4 MCW][~p$|Ek88~Ek8}EkG9B-T,)߃ϧK9jdٖ7?WRjJOlDW?Yj`_H=^ O褃#ACME{i@9D!j$rkȹlWKS1\9B n4jQu:?hS5B]a +5Hl%IYCjuz2yZC!4 &^M'čZV/ӥ"yDpM_A/d˛Ws8߯[t`lml}u҈YŖ'8 ɪuvUV 3yM9&MޔN73DfĜ8N툐?7βtn%/r"e0){Ѭ'I F?ˉQB2&=H<מٍ^)2?c6("OH[q- AU?.!.d(p9=9h{7Lu+zow c"vD4LDB+KqT *L$Z}i:nW^TYj\ @dvZ7}}W]ac|@ٜ>_o>!C.W<uNõyFc5>w$870tdXf:}dZg'soΏd&Dc:Px}μᎏ动EmgՕ3FCoߏuÑwM5wի]p^y`·J;'ty!FCZh!(ٶ^*MAIlB,HCr%LDD*wJV(" K q㐔-r2ݶׯ6f\i/{x`c{5zH.5BdwjR Qg؅|q`|ʒ|,hH'n zbMn`zch}O4Ҕ戛mW $P2`U5Ln~15~!9ve8>=0$u"7# (a+ 1d@!ڷAc+t"3QMf+sPCWny@rHO X8"?Q[ KnAZ lCX1ik =vE! wMG*N؉)ڣ3͞.FZ|~^o^VbtՆQMZ(\DPjXڮ.K 1'Jb(HdKo^!ߛaԸt)gƼ_E16f\?·'[{:ڤ!~oK^kw}4;Wo|#XL\eR/R88&U;22V=7,CGԍQ74`?!Ub:"(*>}Qq`9Ռtա=t5rQL5\aԾ!Ivjot'|(jT8~z M #E~QD{WgwEXcX ڋ ,ghZ#ĚCG:d_to }/8U{\y:dr{pSadlfVߛzְѝi0,3:IStz: fuA@bkEWg`K cDj}-Q.N gu n7Vq!hI1ѬdjbP h3$3Ev2QPD1(B0NB1kJ@}e$L;6R!Y{=MsG*0&lH>Lpۑ 7y| R5 |~܁5aJ98Rdm~GMbFޟmBKq3ʍ8DFY^J`44Uel!"27_n>>ͧWoE~.KJ:fұeRøq*6gg#3qu(g>ψт8RI09DP :? 8N|V'4?S8P XVUtpF:Ipq{W~wȁw<%RD$|UST/OuA<9^]+ Ǘ?̝v3%tQIQW b"%>1.VX4WpHQJl$ 4ҕԆt$,Bju㮥љOT5_S} 8CV!* 9O,UuQ -]2Qb(ƐTRk9̇:<І'zQ]nR/'~^l|VL羬h.ӊ[(rz{eW'}񏟢1>i^~ľ>OEGI7l1 h@O>m"S%YҋjY5$g\};>^%~KUf1MH-(Ibҧi(}\o1o.E90ѷ xZjqFY~hT^դtDž[,]:~y7cn]zu֥[M~lfcVۊC *jC4q,QĠ(,u >`gq1z={d]>3 e[LlO#BqT-%Xi,>LqٕMoo&/Jl:_g&eȈ9'˭b~W- g5FX0XȍURX\0\wlV܇`l,CX~^@Cd9q-/QxWB Pwjpm] /֯^ 'X;qB$-ʒYѸ-$( HCIwTJ(3 XT\xCE;'XCEE3x·];'w"[KnNd|׋ZTUe/~y~_ ^BɢXqDUUQ# [gU Qe1C Tž9tt@׼[(/9LƝW ;H dG0 AuI9+ ,1JM i9s1#PDCK (o>Ja\vƭTFBZ[`YR--K+_o$줬ԗ>nj=kćIZ,Y^CFK`@J/`F_zW!'Nqq#ϗֽDQH;xڪHz=ih4FIb31B(lU3p fFyѪRH NX'MHaTr>s0p7_L@t2WZ; `3]8]1Π(VɝQ 9#B.*׳Q˳3zX<;hٗQ L4AAQH?U82AQdˎ؜VPƿ^ tS)Ka,Li[)\.e큉y)#{h>aoJHK]6ސ7ZrX{¯9v AD7c3  h/k #URF{]ms6+*Jw]7٭헤qⷵIfS߯A6%(-;/X@4 }.a8si}ў1 aB4I-8yMnV0}mC$uM{]ʳe-][ۨUlSdITe=Dq]T,/Ʃ뭍6=54Hp t6. {agUl*h{®ǭrV޳5IY{=pQ5Ũcjȩ y(.9m dDnMVH ȼ+(K"x9sc ìqfTXO<0rI%[m lgWɄ>+eZƱY>!r1bDrFЃd-X4e @>3Ze[(D@mMÂ2!NdS3Pϊ|[RRN+\<ϹrY+4S69E&ɀ8DSC-lF m'(J2Z=%|6nJ]SV `# ɹLRkeq wJr06 CX?WnzV&ȴ@3"spd\!ׄX``KCRVβA`_|޼^U'v8nEӭܝŜȋ2eAnunr^ 4#ޢa1˵0[fk?if!@!$F?mJU YSP(Rs:^\#BpԀzW68)KwysB+{JE/Aecq0O֊]=%l -)CG*W)ޡ~~͖/>\_6BQx%W qs'JG(=`GU {Kfj+:Kαor>F?qazZ7g_=;XTjf8tx02;T\L~㙕0=Vb^oocChEzp)'c&28tH>j-SL9;3'Ěxٔ-mbFQZ(D7GPJI5u-=T; x'^<|!y]u@av^Ţ9H~q' kj^TZL5-I˫՞%*)IGGgkZ)9xIԆt]U8.*bWm٨p eㅗDq7{ RB'~ !F2*w?!L1_"d//2ċL燱>ac}uZ.n?bfq23 ,'ZqƆ?%j./'=CI9r>ybHŵ8I'pw=&4YEGx'`VcΏ1QER o"T)++@In =WpJ8Y r<) 9/sDbhfJjZY!|s1Y3 `1T ,8t( 9~.1`cX'Q&1M)eMg'0AG!̠75 p|#JD9ZM*jS@MH"Bhe(Ptj6Qo _ԥ(գ1%1]FNgDsPm+%u^C:2O]eT[ymeX-`#*o.1轜` NIv);/pIVM>H7W)/:M]斤|@?O5W|*y/>O3Cy͛䚛KA+|6P,OE(v$ƨ9 & bPsoWNSߪ f#'r@q(kZ1Xu~3>, wrtNW'?=J7KR&G=ܔ2h#JW˧gUh(&9a F-AyA{ohpwrr)S8%/QΔ[Bzvݾcd4$a00n<@R"3 -Yisn47Jurr5Fa+'Z|L>עr [msϬF*9rK 3imA-$yeZHaskЪ2w"shKIv:!.U\J^2>(bBq>.Asmw/7;0R%iO5Oo~!UjmZWv7lhxrwt32aoU4E̡m@h˺b1VȏN+EzOG*ht:4䅫hNJXP(3떊Aꤾu;޲8p YdAc[UtCdN;HE DY?Px;n/in~^iN*"FKDmx22yi#P7W6}rƞKSkL-ړJ}>O>/#_Ztq=2=#7=~,ù/dN֟( ІJVqzSLMm '>`wxL0ѽ={]< KSH)˦0!dz5e.,궢5k[^u{?G5_Π̅@5 qסzQJ/: Ȇ4JTOˎ *7<9ޫq;;H[pM/'t5ldT/>m$> ,LѭZ 0TjrRפ"'Sa2ecYtmMWִ-z|RQ*!BTfE;q~ Y>30eNnph,B %͜cQ I"D -F$R* @J"0ZW\ئvDJ(@ m`/./'4ozG}\n'J0\M95TXB - \-g+\TQC$_gr,D/B2_ЃL˪={u>(_] ɲڟB5W+ ^YZKBE[Rk;+>LE~q 8yHIJ %>C}(W;DJv΢R4ENQ)z1巾,8C>o>* :Ԧl-=/g80;WWV^pFfy\F4`bˣ{|$jlY2|=BӪ[̻_a[`@դKdYՑ„t- |0ux!vrJfi)? j#PjzH_ `@í ӍfeIVbʠ#(tß{%Qx7c̐SbS+i Ě͇Ig]69b@ rxј5+<⢧VGM9ʷv=,.G::G FG90)7iMs*G'X\"-&㢰knk6$iO/-äRbkjf٫,*B`YKCu( ^f:0ʶWpx s5ht={(%h7A -,z%G+4w=#9n^uxoi-<#Jg$ė犠>w>3IP6&2DuT[QaV< AYl4{Ev_j u\W2 jr0v@1 B)xU]QsCw|$' PdA4ThFnݤ9d~utnݦKqP^JG5d7hsQi }Z 97:9hg}HG!_T9C*#;ϼy. @Gm.FU:*8ܢt6S޲dP&ZjX$H5D%N|GGfɫh Wu Bƣ=8*p+#;{< ps8" lAad+GEaLbX׋d(l,;8[k1."p[("pe\;49b@e^zͶ^1>pim>EWu8VPy[MPua۸]cZhHKwYҿm+RU\mîLeSS hƒ_*5(iD=lKFhtӂọlh(%K!/j39hטȣq Q2]Ći='1 9M'Yy^23`R9"is+HbTV;Ŝl e%GqIfhH+}_ŸR%٭3xAjZ 5RiLT[pFro&w*g3 2));۵+7}/'jNyjBtAբh@m8+Ձ|v*=gH Hv檱azlE11EhM!:֡T1;eTr#:+htǂ-I dHĒj mQ! e}wLXCcb -q8ԢoG17ApWϐc,wg7zT%m["H;56!ҋ| KZ*ڃ6БniiLOVԵ4nՊtw;Zꈠh09QU\(nyVA(9ھl{΢$x!%UK@RJxHJecZYMa?ŏ{?ANv#qC3C&JNyeQy);@e( 9OMQZn{VrqRx< E'SD\%lDȑvƶT!BZ ٓ `{1Ti` 2n- s0cn} dJ7zk uEK|7l6jN@(Mf?^DQڹ'-U1bbF'}lX߃? S΋c#O>CW7M1+.VjM߀j^pwo#pvQrbJG4oT@a~+WϢyedJᏓL,I)Y-4F6[xQ+|O5A$a]=ڵVBf q+Oxļ5;N菔2)ěc4trA7qj{ Z7#J":/(Z_t{]lATlyDoavf@ج,!~J7Li)])jxTOT20NEnO1lX#͡Ɩ"W Lg=.fvÞ{1[xd>?e̷zYu?1𕹷Vx ]|)+?x?ftG.*OĂp}siթQ/ YxE9[_3hNqُs:MR-Չ}G֡Lw4QmVZ&4+W((;MAľu;av,${!hݚА\EtX߭]Vn~BAľu;)TvgAZ&4+WR>M@X%Fs(bR-e?7;$gwD`$ew590R&|>O JB# rӑUl0|7_Thi SγDe֧v4тD;ezI9T[5o/^ZEtW)AL!uՀun\\}S;](vH>H8x>~(|p[_׬yh^_]2(^מ Wt|w  3V!#42uiDda# Q)_ MYw<f܁-/\V\jç(G;GgX(-ޭ>*vB= wQAVhb0A9KL$f 2K^h!?,Oq|*{Ŗ#_NGsHM˂Xӛ?P!M(R\ -Zpض?lp"-rwX&f{mv+W`jqwXswa̮sn(D$.-9Øwy(dY.f\wT['2gQ+R V8k,҂e2sJ!NO0R6bȱ!diԚM>2!ݎkP+}iv@bXvh˳1e1kJnB1d+ HSDa%lpy |U'$*pUNOGlʝo ef(3)Vj0GHb@SLiGldN(qb ,WN"*- raUC:JЧ _ji2iJ]:W ۯ +scf'TAۀŐxPOJqǿƭZYǡ=־{5Ca];_wsln#cW4&5A +kȕhOf@}6{1ڄn?VE)ß}dyT {]u76[)/10Vq8d['Ux9fOdx.[OR^\8$gT(g>߲Dc-V4>ܹ0RqH <$gTJ;Uw^4t-BIM{!\vںK"ĩꕴvwRlCDRt@ҋVob-߿Z_qʋE/P$** 5BV14*89QL]\T4.-:^d9He"8 4O54!LD*g9PXyk 5MKθ(qd &)$a"sgαސWܟ8E3oh"Jx,ָ͞>Cf?AY6`ΒTL'_S>| lf8m|H=_>1Ɣ"O,1*37 eGf@جr oԊ|&*ih Lڝ- `:/O>LS OFqdZd &*23#̓w2BQ-eŋO $SYejHb c$N+&opk( Xb-6 f۔[)U `RM|y.}yk&nzE^mKQQ9Ѭ4սi_<uQӵ!/P ֌K{#Ԋ.R\bjsJZ+kЎIBQ)4yyI9*7YRCĪCr4gq#fZIRS J-EVY-uXa$V63nISL+H֠Wr{SQzwO.&d.+1ҙ~Z/֐=)I1㉢^Fv S{7ͧb}<5y2Zfsi2M$ TE:2RJsFJZp`59rS3,T1"tC`t|D\+lU~uiA=x*2FP;Am mQ7ofh}`ԙf-i[䬽RƎ#;kXKZ5k ړ%W0?hOGc>U$#U΃vAUmjPp3 /׶%B^nٜ2E_0~\Az|J٭X -, ZUB[u%$]>zc\{ts? YA/{b"8&H'' lc%$g7OemYbe)`VYXU QUwJ$#|se!T2˧;״QP/dWdǠYɆL2 =d@G%X^xy!3c2֙Nޘ3Al\~:mZH>nJO 4Nd>CRڍ_ _T_ɾWW30"so.5<9%qo]f9 Gf 4c 7R TPJKo}A2y TCFqZ7Kih`5ۗ8(k 1cqM\ʽj b)MҟO+tZg@I}#{ YZ+yɞ]d|wG*Nj\Aƨbo_Vy ?rK}S$bןy|=᫏KM_d)ݭWwL1Bj+'cY]>g}>Wܥ~QrY!% 9r͑)Lm 1RCn:upշqN3?U%pA=|?Wo Z#M9$#M9Ҥ~T!{dQΗ& [o")Q+}&ᲷӨ Ͽ>rII{ĽI{^S݈i+6aXr̥aK9J'O8֞n=:KH0j618ȂLl(U^?% NX5Oa$\ IwEmY@6U޷yJg˂$(-5塢n }= }]UgBbDj$2Z3Y3I"|X ~$uRԙAwWn/[~bJ}D KQ겅%(`.I-.6p^']-Z]a/{hVOk F),+rz5 %o-ז1 d ՞:,R# {d~yb 9ٚaQ]$=pjg4;@Ve@N BI:'#"&}d8pp6b8 zFbxە,bv)Rh٬(^; )gSA&_ a !o4mK ڝO#B`*6w?=>EgyNNZӈ \pZQDd8F]L^qKΨb(G"*)ȓ_ED 4b@ukTqo{KOpP"D9NoqIg>/tbZo"UGyx,Qa흙\o{Nf;pj%mn(zN3V07=9x %3EzciH8wL[qT sUBL^UiDA&'["سcuR-)=e)S0]aVEi'H_*I}4 52DCODwI ɩw*#@Du5KJZ:[xkI i8!i@;EQWEmN;Y8'N P`bhN{5xpbSpXF}-aG+ˆ#vᴓMImB=EI4V;HN{,x `=;pP*@X$x\ nx!lUy //lI*b͚a4}O3F`*hfJJV0XU07c [A2,9z9]@vtW+fCj^jq0I`44ǦSKjrJ5Ӯp0Dq|D7"aُ*jߛM#wH2DHC޳ͱH S o7Kqmw2B$'އ8*KԻ:}\l $Ր’b1*%IMT#!Pn/ƓPP ɾ+|d<Evn|Ӓ9["; #NWBRPlpGGw W`#Z .}5:;[ !F(җhTqp 9hd<32wN8SX =1E2HcH\G1FCa NIca·1H0GJAHP(^U3=WGxx5N 8-s N;fo1XM&rX=Zt"pkW(qRAD-||G {PTn@|AUh] 5sTSDUoR*6&$wԋեZ Ku6ҕ a[ŝg"HL bWz]rKV:w,|gR^:\"Z -P"n{)o??3@/~$_xq\+q~y$|.0 :y%+K,d@XX!  "J%8a A B +aF8` +a󼴆3du)CKZYe`d`Cg iǣ0.`Lp18ϱOWI`%`q;c+y0M%<̞qe[)q,SK% {&{\r'5B~,Tۉ |1?gfJEf KRJy $Y'ƖB(2XZ:eU:\7kԕm\g`Y a9K-"GB %B_S\TFW߹2ŗӛ2;nb\aʏ]_H$ #=^>~7@'tw!sԏl¶Id,x?bM7N`UcQzd㦗r{Ig}ꢖ7Ա*z 1\%#BQ)7GF)E#7j#CU v$'G~6A SńRX~BCRD!gLPXK.so4vz0H;ڝvEOA}*K4 ]c&PdG d'ϨR-.{"@iƳk õ)/{xu_ߎPnSwniXkO.G_O/KFnWR\-.ǁϞx= \Z6vdV o5hjAy R` k(H` b"(fyЬaxU58Wi,'[DQ,wێje&[U}sy+3xH- `r閣n$iE=E^Y3{<3-e{.ψJ7N)(M&k.𐂫k-yAuN?hKTO(m2?~|/R~񋻫̼4~U~q9 h/) (4;3 Q9eqyS^gI1s%.0q`J lH)B-ʒZT0TdKpJoP% {(w{4 su91,͢eh^/DzY#d8"vR^U?ihpZK?4΋+) +\d B` y(ӊ~q\CcbVJ.'Q.'Q.'Q.'ugsAs!ᄳSkK3 [d@2 NN'L (++XX*ʅrƺ,ȂZ3T*6`z$)Vb/(`9D:on0-+*߃)@g1{TAh$e&01Rb5`w0N.Kcq#fs͉51YP) р T[we=nI̐}4 ٞKIlYd7W3YYU$-`/"#"3p3Ehe&/YtcnY0`3biW u1y HPe*RN1VS[RU5f=I]*\5 ~H7_<a5W(?tNq"8Vo mb%⃧\ÕTeuh @Oc8=qʂ@A ySAoNu2=f B:x-8 7õ\MY)tj<4Lز5Vl˄-[ƕbͨM 'RprG-i Qz@/Ih%Ի*ӑRŗ f E.­'Y(J-v{_qG~}v7MtFY|e/DO$渾._/MZTZ,mq h?ự1sK%*k GZ)b- ƒ5:E &eh LG}A뫚03`Eȁ: %+WFkXp0iXJQT( ̑`4jӊpȏVo";~Gt6~t6X],TywȜz`ETxbC%C6O6 dDZԒY)[s\Tf1~rO+6?[ĩքPpǓsA(L!KIRZ=*КT3DQ;o`뽎'R3!6\M |,Ϛza:&&5w{*9K%\Zz:^LƠ\f_lk[ƱՍ''^);76e w'ݷ>7ɾ@[lVn,WZ MNyYjPDO_'acJ$FH% I,x.%2xh6fUu؟0{;NMHbŽ_.Uұ쉺hGrEZilx6s[B+Q8puFבHy> <-u3%WriZ,Wr jEu[P_Ń˶4 yЈ6.X5Ѯ"/^g'HUdgiѸ<_7{ޔn[*sE|=_M@^J;?2fRŖ~6?Ml*KMc);/<DI#)Ak0FQ2VǞh͌Er܁.?lZ"t.DJdNa9 i<)t(bResDrLW;,]Ҕ;_DpߛL?M^,Gyd}1ŗbnh^,~NR ˱7'K`R;0D7 ]KR%bTlc**?o0XCT EKa ۴#Xѱ h<#wb{يD0c*.9 rtD|2G&n<b C(T"rP4gDTJV<U@D(q쵶m+P0t|^8B=n9#kȕѿ[Iu3]j{W7&e$M[bq5'H6g$Y挓,HspH4gr\oYGLXdkÔs:n}|  ~5S+tw֓mɛN7=}sq`},m:Ɵ}q7730W.hq/A&`hU;hEpŋa>ٴN'_@7ASGmؑVΫE$`$%/Xܯ%o~bq1L}[ tC,kea2)]CxV]Q׆O?*G\pYKAk^,VD5+ |5zTH} 郄Ⱦ=A*ck{x;$*7}ݩ Uf˕GxDA[+oj&8ke-kv1ψ.ܑq#c`~uFq 6.۸n.j܂ *hBKX`,JGX8ǃq e-!8Qanz0E梸~hDV?s+X!MFAus _0Rh9&<|9Z2̵D0+(QZȆqO ?2+ǐ)2o'+ ΌRa00OF}Xi#apa,,53nqq~*yX#FF;#Tq,*^-gҢ.3Bpl$v,^y s5#yTS~ȏld45ʓ]TؽVMjIunu98#^D)iXB#!hkEGB(ƚ/rUk(s\ ':UoHsXT:5Uje@R+TI*h9zp*{ߤA?tfvCydmdmdmdmՠƵAQ ˨Fx]?zZXŝ ^J@R+<FjV0K(+'wtaC>8;R8([!Y䊍)(>?PD?e٦dOS  k?Շ8RˢӯeV5X/|_N ZbZ2GLP83ǣ O>>>Rõ:?ǻ@T=t1{o %ф na<ޤcOxCKqO税=T\+/.LhLƽĕ{A;ͯ[Dp eY' }"Dpk23G)_yJL]E%=Vivt*`0[o?cFmSW>ZK b.-L,jӌ1ܒ \c9$vg@:!Dfc%)?-\{&)4#ٙ uMTjNRnPk**Ffpui[q C;uvåܬdVZh|΄`${J"O&bX;HDEʈ;@0'"80\F0k+{J aHRg}pN E b f(ԥ{Y Ǻ ǐͰdg|.~T837~:O:.y"OwR|!8fPONRE驚_Ϳ? zܠ*GBA!ށiyмRPpXc TpAX 3T?ޓ`3!,<M/,9\0%de"- Q%S+\*÷_!SΪ~Rk,f^=bG}bowjʴ"pLyQk7%M})8FrU":k^>/''WPUxL<`8 l`iz\Wm0S(r q!ZVӜzE|*XRǗeelYoYZʸΛ{{DήW^:,N¿A+l ~yÛmo81V;ngGJgYDV*u`nD}2|}wB?yc$,Rv{Jp [.);1tWiB-avkBBq͑){[ Ab":cN<լ3=Tք|"Z["esP~7w`) &/'/B z$Ծˁ |&A:l=#jU=ʌUa] 9l[A'h%an+ '<܃ifW?)1|UD#?Z\'^#l64'TNŒ͌,T wȒS}ʩW*#)R$٢1o-G6Xhb@rpƭz'ӯU)g;_A_|ٜTr4O$c\0\ z5:x.St$HiIgFRy!ſػ8n,W-0bb] H݁--iR?d-,M ,V}29Su|)F\wZō,^L z.wl}CY}11!E= igʹp#%} `‰,Ґ-#=`uЃɇN^$a/fjoD' 򧖍&zvyXCyL7bK)KX t3x"5BYOaL R$LB@ˈ83-;lxeL+'&!:;$̄W)IqphQbi NHDpE,[Fc {{F"ϵ.RSOYP ݎi4!^ ҃"׭ zp #m b tx^<9=%9UOaYOAFRR%t@K*)0Xxtyyǎ2ؤ>ldԄ*P:ibGVc2 P [Ndv\HǠvq^ l+VJj'+և C4P):!xWd,s^6L~3; `ER2: L WF0(,lґ#,<>R:u-Hp N?JQ n笏b$ DJe,FlKW0LC(Y1h&i1ٴG{~X}tZBIJa/Hq>B*I\a9 t5"4w$TÄg؉G oK1\2Dg|HW F1ʤOc/I߿2(3k57\K"1€p4a* ʶ\wc*9 AĤ~<%9@uva$xaZyqB}Ao||be@,CnO_[/&RRt6 -EMOHqj5Ahq5͏)gyK'yK'&Xf]zxjMm\/a.̂V7aTxּkt?shPD=޻LJLJ=Ip2V  .a0fj}`.[ԗ7 |K0&>G խd-6iOֿ=ri )j^70d* "U+*R+bFwmۼuwSPS4H-#`HS #h?*%:nbv&DD Hɉ"JY4e3=0aeTq+6a+ U1?4#]uۍpfSh jVobb"Νf4A<(ʔHttA 1!chKb qlI1@lӅ'/nIai{}RIs<`g 9CgN(Hݶ\vۻ1Top0UOMJ+>BNE"(R;%if%,Eo˳WHRX%B(" Ʊ&3XDD >gHp0\ QFE2xꎊI#P)_&=.g?+9 L2ͮGE,bSS$^ĝ|f.Sk}etMջ}c*ƞyA^Yc{OBM0*Cf9X:lVFPːc#A94BB,4Ǒ y5 Lg_vuVNŗEl:1=4#T$Th SPSt!F4 V[I\ 7QUVc %) mPxY31-9:D D8ƞtIBg`o~Űwfqqi=1Ng0[$#%mas4&9،fnQ0TuQb;ƖBldv f)FO$F?[RXVA`Vp{A; gh 6>H_ __ O B.KH r dCT"044 ]I5!۝1p ]Lv_'(i͏}Q%t!^I{a.P$]ӄKlډY{HE:_a}rh侣R_$!%C%'3yKLwWI'L<1;&֭''C9&f09p"ދD/?O*Hċ"r I4BF'aBĜEQqbx/)æ0:\or fR̒e-@LX1,2I~'()w @ڟ?Kcu{91B{ tVI"A,pPW'4 -X!z+,GbKzͥ|ݎLа~A`4+Y)y )AصO9ƱR61q)[) bKxF4\a\n%0BBs-)zj7 ^DR RD9mF p*nQ m E@8b,{Ѯ\ZVWJ6ǂ v 'PfJcF)qh[ "TfRTFKPq" ~_G̢ 5=U ׄCKd42&7XG>(s1cLA={ȇVxPՋ6ߋ^lK6wgC.D¿ÌV%K8}⾼ęmW~{~%,jx<#*h+@EfA B*&~  sSZDJp H =,Fi HOû{5 =L0S\YLc* Xg%eFLo&C^#:In\ ws1&#|[[?SiO2_mwg]N[_ݥAiEDUWRqTYɁAs?/ai3Qv1!Iu|#Obdّ5b#gotTV/|\&/om@ E=P e31 篃[cL\a ($)%@`<3"ΰJ(+M+ :4 iJxiALA3SW9$ c@ BHX*I9et^R*c 9̃S2ja7L\a1A)kFMIePQYeDJB*eS.`(M„?̠G]c>6Ǭ.2t;^_)C"Y }PWZ^1Z[g!eU&ͽZWc}WzJIQ1P:s=+rPC3hX-἖ U31 W yzZŤeH^@y{3OW3 K Nzⱳ"9 D@*Q|}i;ºX2PH{ c.aLu0%aį a^긹'Ӑ0sBLj!S$`G1aG!C(E^pq(X"f `:AP^9mr>~8)*n6Wa]-HF Q݀/7z\=-qovqwm_w&Xmac`=t ??\`Lp 79IQ(#?PG/0.]S1΍㢎׻ޓ0ik2c0#ۢ,0(msʞ ܹP1 ,aie*ؔDh:: ̣vo$ui LE:a+ w֝Núkr;+󏛀@o\3T=!5+$8 rRS*ŀ9)Hn2Q,@tz0w~!WIw|cѐ4]=d1M٥7Mj=@d뜫U \F5H>cI1Q ^|kQA.GQ^(P0 qLpe 3Ne*|zS B1uΉf'Q~є.0ƙ4Ox|m=9ݘfJ[Ԝ;<}G-G]xF"Ke_KYlDA܎h4;.CX18y:zVnOr;9* d;= <<7~r4E?L~77""b(` m=ޕ5q$鿂 }83k~CQ@4]8,FuYYYUyHK"w)E R=GƵuoFr m P!rkX6-[kz%z-,kJ<ȥ.rQ@ĆF`ǽ<-ʬU i rYV*|jfWYrP z4CJ$̝ne7ѫJΟ*q檑gg8+ev\i,bk `b_H?a:{lJiƏL8e=id..!.ϐ &pXq QJjM^3E#PZz'-bL@f,V df5˯d.H*S_SlLI^j71/ۥH@H"TBD*"ղT$bat40:,T>(&ƫToR,h21G(A^nmJ](6:<[1= U5a7'j~ހٜ>,Jt߈?)XHj`7˫&%Mn+]qMdX-ϐV'O+U !*ljїE9ՓC}It :T>Or&0Z @|AM@TuG`L4s U[d +{C9XWy1W/n͈6Qҿ8e|:k=S9PR-5#M`RA"EoH?_+ݧQ!&wu 52mZ Ӫ5C]AHQ:߫@[Lh)ڲuveyNxƔW `a ܧt8nf\0wj|zF-!ck򏏝kw{^m˷]닇"\-bѻxy[e|b!z-6A-_iQIbUrtLK}BcJ$~-;؃Dij|طDNrã;hxIs$!CRDc7={1nbׂstXsqY$\>ٷ|?6ۅ칝 ؃w|)(da^7v癷7]}~AZ[uvs+ X磰s^j^a,/{e{DN/~{G|mo0 6rPvqów7-\^m{m$^}%ŤJZN+ci˿7ǔ tz[XxNjC 9ps!+i|:Ky`Hsv3 '|ހQmXX?Inc\aˈ aOQt:Ho= v")IV^a5A> EN]L ǑE-!4X&K*}|C# ŕVJvܫ\xMyĂ7Poqٔ# L@$1A;n cQ&DJ}N5ce56yRaB!HQQάR-b%fOIpY rezr_n󀚚reb_Na d#wvQ/pYVPg,GK+c˥t ʸOoe(kU`aX"` 2Rle |/𧠤e3eck6D 3r;kٺsf/tj+}X\):hR\i-ȴ:NLBCr)XnrMG떋A}G?d*wX<Ѻ֭ U4K=x~ͺ ^ߣuŠ侣uߎ-tokݺА\Estu)D'g8x8{2ID`u\\9-fQ-+ezz9p`f|C_0%ugR+ [PmbUR#w><0E vB )CL5#ʉ^'$Ѣڄ/7AOxu5;) . [~{p}"9i-c Wf:"|c k,RE!pp+1yZN)JQP9ƔQN3uc5Ս㛈-QdeūUY\PdΪ(u5q7/zT|%U1 &tꖧͶ(O_L9ͿYPJXultjo~] t:ps]&|uNF,yYBDA(Y -.I0&_MDB؇Hڳ.ZW@$t=̘'B; +R|JrAJK!МlnEFRVjwYпӞW-.P ޤ)twy MY$㭹Ghڔ^Dlm>UŌ$g]V+sXP؞6mt\PLjs˿.g:@)-ssm\u]Z=©kX0/,ar{_v6O~m},QL`G_b!"8$ Ŧ<ֶ]Rȶ.Y6xE zrwjIh;WH_CXmJB&LSY{ivƼaYLU?}N$(ඨ>3Z)p@jPz3WG<-GNco^ Ti=\`3&Z<'@dD$m"!5|5X;-%)b2$#Ha}Fh QeRI^3i#}bY*"@uی<H*4ڃ^!`tHP4$F*j hA/YH =KZћ-sJ`ɉ)-E9k5ĕ08, NȄ4gP :I 2F#G8ƞxPZA|j K':sU W& 9x"Z Fi 8xA$YP S= <M L4f@ #fb=H[W̱q`|a$Dь *FYsۭF  ZC6pZD`^JDX$<`?*D c%>l!#8n6Gpz kT9Kt-DB $pdSHp^!18TF0ǤQB9Jx7*@Tʚ;=7ߟWbn s=yX0 Π5JB47Dv*):jRzU챒$v)[`9cOL`aevydS $ێj ()x:l`2,eL/rD̠QՂ"Q 1Ke3Ĺr2-|fYW[sIԿQhGI81g~v[v? _1xG'lHmCy^ &xWg|%ó SS?f.$R+g][o+F^v0x)Xe >R$ˎ؞ߢlDz-Yb]W@7hv_~ \mj 짅9!eT[9/|kXu[2O@"YJZ~asd$x67^;!4NF [Ӑ =|NYGʓJJJPFEC֑/3S\G>wic1։ds"DFTZ(Ke|#i3x沗&d>o˧d '{d:\φWƫRJ=kp q`^cׅ@%S”Y@ͫ [yU ޟkhUe<Ųli}7}uq?q>JDҒZbYwZߒ(c%42CB|ʆ{* eG672J=j=<`|ߕ?̎_&SϪ JnxU^3ȍCrߣ @x*ID]X}@c*ak21{$.!q:{r|;QQ278)^b|mLt~]PO1йZm4eC hplNP^ }?Q8gag/ ort2IE0Ԛ~/x #>vpQ(" kg0F _A*odsi$5}~u nnu3adƒ`IP 2qB`d,nT o,TN@`2cLq(]KXĎdoEx1Nfb_IE|cDk!}K$n j1x]e(e<9 ï}^f#`9|ӥi2p&蜪@;]NXj{$xPJDdl\nŚ'2FϸmgQmuU9lY% jM`3dq5c Sf|V1BbH)h+#ΕnN~ZZ;3=7~;Gʓȕ]"g} 8~N`dcokwe$2]]bE`ɗTs iu t+Uq|sE7nh+?(׃Gكh`_.YG.VAy6w(rxF% > OW75 "t-ӳ]Ebo(iR0v(_UbݭM&qQ%~1Ds(sl\B`Xu"Yў%/s,Ƴ'[Z޶]@>dsӘe*HX񐳠XG֡rmMO\rJNVe' t:툒fgByNq+8ƳsHBvCaרtwWeb<UI\,W]UE A(58N0ۼqlx'+)^x8`ɜ)-Y]"搟õjpa'Oj[1HˋG&)#Eseo wvύl{smݚ-kS?6 0P4KR'"rɕmiweY^sm<磨~?Mid]]]wE9'o|?7Hi<~$؏%V :OlϞc}zi^Ӄ rQe;ƴ(+WG'k ﱽնG|ND]di''Ҝ.Yjң8+֠ło`,?V]yf|y>6jV᪹g'݃ We#O-PnRv_Z18ܙ$#7e=D9N4a58.sYv9QwA$ӡhVuRG÷ړ.`Đ_[COkW8 % ' sƺz*c&m!,'\;.f7#lVl8rPWxvdqL:gkʻ5h˯1W 睱Z2^6!ZFR)BH_XF͌~[jʅX9׋x E-exZc<:5ʥIhۥG7:ҡ$i1$ V<00@ƑUR. UFG6Q%5X$N@j|,eZ?tSt`2k3y5 ޏbQDsms6ijcq%j}I/[5)mՁJIjl2&Ƽir(-:P$'S& SkHWL9VʊU[t;Zkҥʚy6A0*ufYrJT`䠂%1#8kA[>ITp^Y ⡳dL5p, 4˙hLQP9mfE3 3YcQcaHjI&6/ RUg63G0GBS_Gg$#F䁞`p!ԡZg a/*s6P /+ٙ礖L2I %- Z^ ׽@4G+>q{N8NjzdB!UcBoXҟ8s[,6udUppEN!1?-&uGE,YyZmI|bطlvqqCVH 'pJƉߣ mL "uN?1ާ/rkۯ=ǩ~Je^+Xh:s )pZȏ%B)')oK5 l8z~dM|rX_4=5 <{ qccax 1O!_LHyia\͖J5'#<0OD7W+߬ rB+=RD\56Y z7Yڤ@/\!a$_c;>!Q+́=o]t['BHuQBȵDx`hLUt 03>~t)Q-Q{9akOKd `'[,q|Zp>&o]$fyg;^gFv0yJ[k=y|u!_NKaR ΧT,4TQ{JSsA׵'p7%eZd䜴 ňNKcDzHF/O'6ll)#e}%!ܶ] jmv:?L3.*,m܍D+;QKQsy<6(iNJ=R'|~,kcdUͻP 5 e@UCp}h S]QUR11OK rl')\Ek*qRDovRFUSdmfW,%yPӷt^0z 0łR ^%NPkj (]cqH{{R( evpXa$֮2p\clqaҪhM[PA?igZ`a-ߔZ"TdQ9m07ǣjhٻƎWbKbY=6`2ɼ@kbI_d˲-G>Rc#Y$Er2L{X_[@ $j7k#wbGq LsZ/% PFvVwiG53,xa-ܔ2 mj ]*C[o~}BXk9>ċ+78Gp EDM ʃYO (>JT#>$ʁ)#s*&h伙 Bʗѹ3#\ae*DZ!rn5ڡFG5h^jimBn^xoF =O6z߹h' m)gIvS}"EVޓW!Tk_z8e|PCԎH,xk #c18u˩z%L:ں hl4WegjGyurނjG̶/:Z)4b+'s- hR")Fri'l!x^h}zT!1irӂg H. QЃEDVjM}mMՙG^;G:ׄ%)[$I$'_C(OEr-VHxlG)ĽLfYqT643ɖxMk]YH9J3ۖ& WTsx稾ynkNfZL%qoX׬q^WfM =+C!u3e ^\_#"G>IcߑOQ5:߻Kr[ ȵ*ӏy~gP-TD:m-8k{7%,wӂŜhLuV\& D ts\yF^\N|ի>9NI<ԺEo<@#E̱|`'5o%p[n-#@ ȵ LFy:^_h![5>?;b9Owo:h2~t}o|w wZT񯗹ΜX;#2χˇOŰ8G![fSLe.X3N+/Q[ ؽ"^MRݍez-aC-n&{bCn]blƌ8=ωPWS3 >/^R<zn@:e:  ֆtٔ4譧+j?\ЊZa9p_оpCc9}޹q x8>2'k?╛'\d><ɟgs|6':؇of[A+r5xbu~hA=>z2}TɭL)L)0DP#Em"ʈ.yWOB- ijd!ãePrxy|u3LvN\>R~BTd8E̴,"d9a>KBOk{_jрѦG33AK"S$9.EC!R 栭ƚN2+z&AqK~ƃ5Y U'^ &EL֛R < ,W;kM{e#p+NEBBY_^iYjt2A Aڂչv \)V8|]A{>}Ć(Yrlx/B#===ë큍{\7e7m[ܮW=;Ll|:;:;Mdt\‰g$ ;U'˖iP͙#%J|7S}jbࡡȞ6ZKV;s:榵FocZ0izkmMkmA0rz'X*k%t|y3EY@Z$_bQrH&#,<]n ;4cR4\v!iT=L K!9#|-/6nH29zp5!@!tu m\FqSVXcx  UPIsFta&eFXh$pDFr^ܩ#B'Oz1tBT[1"ѐpdQP(y %EnT7R.F!iOvL2\h-L N3Fh ]K]W#Ơ FX (15$ܰ{FtA${'/z@P< OܣJ:(EP:1jZ_(0JBzz΃v O̎gi 0r؂ȭ#@Ԩ}Łp }7OzBc|^iu;c󹍿$&yL7%,MNܠdUqk,"؝[ Z\}EҵBj)&_tiekʆ}$ xkM<?LN͡s(Lz`U" Jw Q} 4wڢEy=/]zx;$o 3>%A)ׂ;t6+4-9zk?9.8gӌ|j$C\˒4)9 9CA% M#EzD?飿p6qc" g*`! O) B6J]3 7Gk}&p- ;#3\>Xw 3J9m_hV @jDv FýǟIŠ)oC@Ŋ:;8=X.;+F9X=BYKJX櫐iU@յ-~%*LnxPzSݸƵAx_U5ReZs;>X;y8gkW0>gZ>0JJg>9;9#$&LY |Ȑ[%6{DކEF@׾y 1ivGG &$-0E&oR4ճL,Y4O /R:BQ5:$ .d'fù3 E;4;HD1`%%d*PgL`\W!-+PYucX T"0%%g^>x7E8%v5? {>0kC% O(~)!Gˌ_XW87f URDh!F7-wmm /sZx Ab,g ɼ ^ز!˙翟bK[lu -Ūb]Hǡ4R40[i 2-8Eet୐he)xP Eb B )ѝj%eVs@;g6` `4Q[H` LK_ZSmݍjST'pcQy"hxvhu,!̻ CpRg"Ju.2Ҋj`h$VCbI"Kh!1bH$qVO=\Jݑj$+XQ R0RcΜEйT$-*AZtFPJuLm zm4NJbTbrT{)i1Pn1 balHw%(h;(MJK5Bɥ/AsEDĢ HP չݐl5dxm+ѿ.usH8seJNyAk|]ok A3vF7X2~}UƷmؙ| &7ӹP^\Ɵ_෴Åg.r3Ws}}&pm8"iRh\ >,~W=Wu*9zđd\V׎gx?b-|(hV-mUIp١vnAk*T-SmQ#|Z CPH{A T ƗZOs1osMu/y`ղ>rZp1Nos ty+x%㵢Qfj&/vKWWXn!WNdd9^R]o?e:"c8xqy4~h+'3Q^bEa)iv:O fk9^i=c7V4ݺVаeJ_K3{h=|h^IqVݾVm)JKGoM^zG!EQ~~5BN#2JiK$h L$ZHByo@ATk}!c&BրQQ&!$|+3}9GRn}AA6+!Ipg!P.URń*#"$"=Ps{H@tN5gҎ}b XA`@XLˊj P$$k%#FT"1k}协72~\Mrmzruf%Js{Qvp^_\y5o'\3xkP1LZK8s)[ӹFȈ_8OFW7?~Y,}bwԪn /[V ̠ z^5fG74zHv, Э1inlƤ 9[#R>x u}f0JaE']j'b_O}Vnzxru5] t!c] jqcp|\Ah^ ]7i 6nsoh5”_@֮S pbhǡ&d^TO@Klf-y]2|{&QaZ**`UUl)=UךBm7곝潼_8m~/q}޼>eL5-OS7q"0Q6dhuv?%tĊZ 45l/ Z8:$s%PF\w>$"-JƉ3[ud/y[:VҪD*h]\g 4%=߽^pzH4ZOJAjF3`fF+h>d'fu+N-_u|Jn2^Y0"4)Jq5RwBLhTCL8+ C걽ugF&fՂ&~fi^m3@Jy@wfIߟm\!ɨd>v %dc+MJn׷ӗ2xy=V>-ewIܿܿ[[nAyY'ޱ,Gȷҝ-21_n)j\2#c]-UI喭N--IsyCF?+n]W3cZ7?ɎoK6l1CE}pqe ֵXw߇&RE-!){rP B>> sQ>(V4gV/S&zL5q ;p-__sZ* =y|sCR(,v0Ӡ2/}x~ 1dian.vJ+|oYI;ĕzA{~T`|c{WZ?e|Ew)qN76ƁؙinSv x{ V[8lI Gv"ٖhf?N CP.l@;`/uc`ZJh+@+'bw5f~뺠k6AHՖ2$#`6NwM'܋ܹ!p58;0Jtˋ(-S)p)U$@ObCXFKVdN1hPcI$NM$Sɸ^*($٦}LQ+t}&DH>|Ji[UpA L{gc5c>=L0?cgo? @QbMśIet/ a"jڗ| &7ӹH +ID<<<|gDHUdpaow(tG:_/?< )PEO'Ϙv<ƾV!؊*_]BN%;swElyq&DTh..AvO f4dKHbPD t&MPJw.ae14@ōpyϖab9&(oF $ofB"HJ}P NdO^%p:WQtLQwvC!\w]wRSrDYQIAK$'Qx B>Hlvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003310663615155231367017715 0ustar rootrootMar 14 09:10:31 crc systemd[1]: Starting Kubernetes Kubelet... Mar 14 09:10:31 crc restorecon[4702]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:31 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 14 09:10:32 crc restorecon[4702]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 14 09:10:33 crc kubenswrapper[4843]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 14 09:10:33 crc kubenswrapper[4843]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 14 09:10:33 crc kubenswrapper[4843]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 14 09:10:33 crc kubenswrapper[4843]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 14 09:10:33 crc kubenswrapper[4843]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 14 09:10:33 crc kubenswrapper[4843]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.084118 4843 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.089637 4843 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090140 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090155 4843 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090167 4843 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090176 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090191 4843 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090204 4843 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090215 4843 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090224 4843 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090233 4843 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090243 4843 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090253 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090262 4843 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090299 4843 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090308 4843 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090317 4843 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090327 4843 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090336 4843 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090344 4843 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090353 4843 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090365 4843 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090375 4843 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090385 4843 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090395 4843 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090404 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090415 4843 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090425 4843 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090435 4843 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090445 4843 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090455 4843 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090464 4843 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090476 4843 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090486 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090494 4843 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090503 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090512 4843 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090524 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090532 4843 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090541 4843 feature_gate.go:330] unrecognized feature gate: Example Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090550 4843 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090558 4843 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090567 4843 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090575 4843 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090584 4843 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090593 4843 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090602 4843 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090611 4843 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090619 4843 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090627 4843 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090636 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090644 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090652 4843 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090661 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090669 4843 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090678 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090688 4843 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090697 4843 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090709 4843 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090717 4843 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090726 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090734 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090743 4843 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090756 4843 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090764 4843 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090776 4843 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090787 4843 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090797 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090806 4843 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090815 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090825 4843 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.090836 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091052 4843 flags.go:64] FLAG: --address="0.0.0.0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091077 4843 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091099 4843 flags.go:64] FLAG: --anonymous-auth="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091114 4843 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091130 4843 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091141 4843 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091157 4843 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091170 4843 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091181 4843 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091191 4843 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091203 4843 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091214 4843 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091224 4843 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091234 4843 flags.go:64] FLAG: --cgroup-root="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091244 4843 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091255 4843 flags.go:64] FLAG: --client-ca-file="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091265 4843 flags.go:64] FLAG: --cloud-config="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091300 4843 flags.go:64] FLAG: --cloud-provider="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091310 4843 flags.go:64] FLAG: --cluster-dns="[]" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091324 4843 flags.go:64] FLAG: --cluster-domain="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091334 4843 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091345 4843 flags.go:64] FLAG: --config-dir="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091355 4843 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091366 4843 flags.go:64] FLAG: --container-log-max-files="5" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091381 4843 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091392 4843 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091402 4843 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091413 4843 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091423 4843 flags.go:64] FLAG: --contention-profiling="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091432 4843 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091442 4843 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091455 4843 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091465 4843 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091480 4843 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091490 4843 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091500 4843 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091510 4843 flags.go:64] FLAG: --enable-load-reader="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091520 4843 flags.go:64] FLAG: --enable-server="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091531 4843 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091545 4843 flags.go:64] FLAG: --event-burst="100" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091555 4843 flags.go:64] FLAG: --event-qps="50" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091566 4843 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091576 4843 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091586 4843 flags.go:64] FLAG: --eviction-hard="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091615 4843 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091625 4843 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091635 4843 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091648 4843 flags.go:64] FLAG: --eviction-soft="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091660 4843 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091670 4843 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091680 4843 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091689 4843 flags.go:64] FLAG: --experimental-mounter-path="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091700 4843 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091709 4843 flags.go:64] FLAG: --fail-swap-on="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091720 4843 flags.go:64] FLAG: --feature-gates="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091733 4843 flags.go:64] FLAG: --file-check-frequency="20s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091743 4843 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091790 4843 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091802 4843 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091813 4843 flags.go:64] FLAG: --healthz-port="10248" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091825 4843 flags.go:64] FLAG: --help="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091835 4843 flags.go:64] FLAG: --hostname-override="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091845 4843 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091856 4843 flags.go:64] FLAG: --http-check-frequency="20s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091866 4843 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091876 4843 flags.go:64] FLAG: --image-credential-provider-config="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091885 4843 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091895 4843 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091905 4843 flags.go:64] FLAG: --image-service-endpoint="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091915 4843 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091927 4843 flags.go:64] FLAG: --kube-api-burst="100" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091937 4843 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091948 4843 flags.go:64] FLAG: --kube-api-qps="50" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091958 4843 flags.go:64] FLAG: --kube-reserved="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091968 4843 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091978 4843 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091989 4843 flags.go:64] FLAG: --kubelet-cgroups="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.091998 4843 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092008 4843 flags.go:64] FLAG: --lock-file="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092018 4843 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092029 4843 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092039 4843 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092058 4843 flags.go:64] FLAG: --log-json-split-stream="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092070 4843 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092080 4843 flags.go:64] FLAG: --log-text-split-stream="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092090 4843 flags.go:64] FLAG: --logging-format="text" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092100 4843 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092111 4843 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092121 4843 flags.go:64] FLAG: --manifest-url="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092131 4843 flags.go:64] FLAG: --manifest-url-header="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092146 4843 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092157 4843 flags.go:64] FLAG: --max-open-files="1000000" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092170 4843 flags.go:64] FLAG: --max-pods="110" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092181 4843 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092191 4843 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092201 4843 flags.go:64] FLAG: --memory-manager-policy="None" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092211 4843 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092221 4843 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092231 4843 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092241 4843 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092296 4843 flags.go:64] FLAG: --node-status-max-images="50" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092306 4843 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092317 4843 flags.go:64] FLAG: --oom-score-adj="-999" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092328 4843 flags.go:64] FLAG: --pod-cidr="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092337 4843 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092353 4843 flags.go:64] FLAG: --pod-manifest-path="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092363 4843 flags.go:64] FLAG: --pod-max-pids="-1" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092375 4843 flags.go:64] FLAG: --pods-per-core="0" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092388 4843 flags.go:64] FLAG: --port="10250" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092411 4843 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092431 4843 flags.go:64] FLAG: --provider-id="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092444 4843 flags.go:64] FLAG: --qos-reserved="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092458 4843 flags.go:64] FLAG: --read-only-port="10255" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092473 4843 flags.go:64] FLAG: --register-node="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092487 4843 flags.go:64] FLAG: --register-schedulable="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092499 4843 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092525 4843 flags.go:64] FLAG: --registry-burst="10" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092538 4843 flags.go:64] FLAG: --registry-qps="5" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092551 4843 flags.go:64] FLAG: --reserved-cpus="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092565 4843 flags.go:64] FLAG: --reserved-memory="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092579 4843 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092590 4843 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092600 4843 flags.go:64] FLAG: --rotate-certificates="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092610 4843 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092718 4843 flags.go:64] FLAG: --runonce="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092733 4843 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092746 4843 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092760 4843 flags.go:64] FLAG: --seccomp-default="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092771 4843 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092783 4843 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092796 4843 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092809 4843 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092821 4843 flags.go:64] FLAG: --storage-driver-password="root" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092830 4843 flags.go:64] FLAG: --storage-driver-secure="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092841 4843 flags.go:64] FLAG: --storage-driver-table="stats" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092852 4843 flags.go:64] FLAG: --storage-driver-user="root" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092863 4843 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092890 4843 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092903 4843 flags.go:64] FLAG: --system-cgroups="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092915 4843 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092940 4843 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092951 4843 flags.go:64] FLAG: --tls-cert-file="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092962 4843 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092980 4843 flags.go:64] FLAG: --tls-min-version="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.092991 4843 flags.go:64] FLAG: --tls-private-key-file="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093003 4843 flags.go:64] FLAG: --topology-manager-policy="none" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093014 4843 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093026 4843 flags.go:64] FLAG: --topology-manager-scope="container" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093038 4843 flags.go:64] FLAG: --v="2" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093054 4843 flags.go:64] FLAG: --version="false" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093085 4843 flags.go:64] FLAG: --vmodule="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093103 4843 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.093116 4843 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093426 4843 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093444 4843 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093456 4843 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093466 4843 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093475 4843 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093484 4843 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093492 4843 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093500 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093508 4843 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093516 4843 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093523 4843 feature_gate.go:330] unrecognized feature gate: Example Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093531 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093538 4843 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093547 4843 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093555 4843 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093562 4843 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093570 4843 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093577 4843 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093585 4843 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093593 4843 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093601 4843 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093609 4843 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093617 4843 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093628 4843 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093652 4843 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093661 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093668 4843 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093676 4843 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093684 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093692 4843 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093700 4843 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093708 4843 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093716 4843 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093724 4843 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093732 4843 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093740 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093748 4843 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093756 4843 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093765 4843 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093776 4843 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093784 4843 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093792 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093800 4843 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093809 4843 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093817 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093826 4843 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093834 4843 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093843 4843 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093856 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093864 4843 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093872 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093879 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093887 4843 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093895 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093905 4843 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093915 4843 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093926 4843 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093934 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093942 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093949 4843 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093957 4843 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093965 4843 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093972 4843 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093980 4843 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093987 4843 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.093995 4843 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.094003 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.094010 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.094017 4843 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.094025 4843 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.094033 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.094059 4843 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.104863 4843 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.104920 4843 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105058 4843 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105077 4843 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105088 4843 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105100 4843 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105109 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105117 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105126 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105134 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105144 4843 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105155 4843 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105166 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105176 4843 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105184 4843 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105193 4843 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105201 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105209 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105217 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105224 4843 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105232 4843 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105241 4843 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105248 4843 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105256 4843 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105264 4843 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105302 4843 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105313 4843 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105323 4843 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105331 4843 feature_gate.go:330] unrecognized feature gate: Example Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105339 4843 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105347 4843 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105355 4843 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105362 4843 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105373 4843 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105382 4843 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105390 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105402 4843 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105410 4843 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105420 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105428 4843 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105439 4843 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105450 4843 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105459 4843 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105468 4843 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105476 4843 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105484 4843 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105492 4843 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105500 4843 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105508 4843 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105515 4843 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105523 4843 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105531 4843 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105539 4843 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105546 4843 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105554 4843 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105561 4843 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105569 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105577 4843 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105584 4843 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105592 4843 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105600 4843 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105607 4843 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105615 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105622 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105630 4843 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105639 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105647 4843 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105655 4843 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105663 4843 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105670 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105677 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105685 4843 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105694 4843 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.105709 4843 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105951 4843 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105967 4843 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105977 4843 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105986 4843 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.105995 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106003 4843 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106011 4843 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106022 4843 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106031 4843 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106041 4843 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106052 4843 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106061 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106070 4843 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106078 4843 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106087 4843 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106095 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106105 4843 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106114 4843 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106123 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106132 4843 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106140 4843 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106148 4843 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106156 4843 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106164 4843 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106173 4843 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106182 4843 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106189 4843 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106197 4843 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106204 4843 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106212 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106220 4843 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106227 4843 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106235 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106242 4843 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106251 4843 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106260 4843 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106268 4843 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106306 4843 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106317 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106326 4843 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106336 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106346 4843 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106355 4843 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106365 4843 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106374 4843 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106382 4843 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106389 4843 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106397 4843 feature_gate.go:330] unrecognized feature gate: Example Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106404 4843 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106412 4843 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106420 4843 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106427 4843 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106435 4843 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106442 4843 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106450 4843 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106458 4843 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106467 4843 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106475 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106483 4843 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106490 4843 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106498 4843 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106506 4843 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106513 4843 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106521 4843 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106528 4843 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106536 4843 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106544 4843 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106551 4843 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106559 4843 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106603 4843 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.106615 4843 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.106628 4843 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.107850 4843 server.go:940] "Client rotation is on, will bootstrap in background" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.114575 4843 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.119482 4843 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.119639 4843 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.122316 4843 server.go:997] "Starting client certificate rotation" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.122371 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.122576 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.147853 4843 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.151400 4843 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.152669 4843 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.173138 4843 log.go:25] "Validated CRI v1 runtime API" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.208038 4843 log.go:25] "Validated CRI v1 image API" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.210565 4843 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.215564 4843 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-14-09-06-37-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.215619 4843 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.237457 4843 manager.go:217] Machine: {Timestamp:2026-03-14 09:10:33.234860633 +0000 UTC m=+0.547471781 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24 BootID:272dcccf-22b1-47a7-8807-43844aeee65b Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:23:41:41 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:23:41:41 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:27:7a:64 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f1:69:be Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:fb:70:a1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b4:31:ab Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8e:4d:69:2f:bc:16 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:76:36:b1:a4:b7:c0 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.237703 4843 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.237904 4843 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239134 4843 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239333 4843 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239365 4843 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239585 4843 topology_manager.go:138] "Creating topology manager with none policy" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239594 4843 container_manager_linux.go:303] "Creating device plugin manager" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239960 4843 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.239992 4843 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.240793 4843 state_mem.go:36] "Initialized new in-memory state store" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.240885 4843 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.244111 4843 kubelet.go:418] "Attempting to sync node with API server" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.244137 4843 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.244177 4843 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.244199 4843 kubelet.go:324] "Adding apiserver pod source" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.244211 4843 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.249054 4843 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.249091 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.249423 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.249220 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.249480 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.250733 4843 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.252706 4843 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254048 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254090 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254106 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254121 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254144 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254162 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254176 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254208 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254226 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254241 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254306 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.254325 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.256383 4843 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.257378 4843 server.go:1280] "Started kubelet" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.258421 4843 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.258387 4843 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.259309 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:33 crc systemd[1]: Started Kubernetes Kubelet. Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.260487 4843 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.263745 4843 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.263815 4843 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.266606 4843 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.266652 4843 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.266835 4843 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.266836 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.266871 4843 server.go:460] "Adding debug handlers to kubelet server" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.267369 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="200ms" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.267909 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.268009 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.268722 4843 factory.go:153] Registering CRI-O factory Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.268919 4843 factory.go:221] Registration of the crio container factory successfully Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.269254 4843 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.269459 4843 factory.go:55] Registering systemd factory Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.269581 4843 factory.go:221] Registration of the systemd container factory successfully Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.269728 4843 factory.go:103] Registering Raw factory Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.269882 4843 manager.go:1196] Started watching for new ooms in manager Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.268694 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.271707 4843 manager.go:319] Starting recovery of all containers Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278587 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278680 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278705 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278730 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278749 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278767 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278788 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278807 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278827 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278848 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278864 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278882 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278901 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278931 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278952 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278969 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.278987 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279005 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279022 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279039 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279053 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279072 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279088 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279104 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279123 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279139 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279157 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279174 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279187 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279202 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279217 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279232 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279246 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279259 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279300 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279320 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279335 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279349 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279363 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279377 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279390 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279405 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279422 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279439 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279459 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279476 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279494 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279511 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279531 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279548 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279677 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279701 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279743 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279764 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279786 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279806 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279826 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279846 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279866 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279885 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279903 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279921 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279947 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279967 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.279989 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280009 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280028 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280051 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280069 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280088 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280108 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280128 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280147 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280165 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280184 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280202 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280219 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280236 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280257 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280299 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280320 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280340 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280412 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280429 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280447 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280463 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280482 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280499 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280519 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280537 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280554 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280572 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280589 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280606 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280627 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280645 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280662 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280679 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280699 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280716 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280736 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.280757 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282692 4843 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282725 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282743 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282768 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282786 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282801 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282819 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282837 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282854 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282870 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282884 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282897 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282909 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282924 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282936 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282952 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282964 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282978 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.282991 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283003 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283015 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283029 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283042 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283056 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283070 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283086 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283099 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283113 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283127 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283140 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283154 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283168 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283181 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283194 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283208 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283245 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283261 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283304 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283318 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283331 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283344 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283357 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283371 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283384 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283398 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283414 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283428 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283441 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283453 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283468 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283481 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283495 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283510 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283524 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283540 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283633 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283649 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283661 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283673 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283687 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283700 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283714 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283726 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283738 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283752 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283765 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283779 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283793 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283807 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283819 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283831 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283844 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283859 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283873 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283886 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283899 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283913 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283925 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283938 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283950 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283963 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283975 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.283989 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284003 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284015 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284026 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284039 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284052 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284065 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284080 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284092 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284103 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284119 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284134 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284150 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284166 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284182 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284198 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284214 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284229 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284251 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284267 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284311 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284328 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284350 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284366 4843 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284381 4843 reconstruct.go:97] "Volume reconstruction finished" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.284392 4843 reconciler.go:26] "Reconciler: start to sync state" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.302959 4843 manager.go:324] Recovery completed Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.314599 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.316311 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.316379 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.316394 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.317407 4843 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.317445 4843 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.317467 4843 state_mem.go:36] "Initialized new in-memory state store" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.335964 4843 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.337548 4843 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.337608 4843 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.337650 4843 kubelet.go:2335] "Starting kubelet main sync loop" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.337718 4843 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.338432 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.338513 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.340384 4843 policy_none.go:49] "None policy: Start" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.341190 4843 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.341218 4843 state_mem.go:35] "Initializing new in-memory state store" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.368100 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.415303 4843 manager.go:334] "Starting Device Plugin manager" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.415696 4843 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.415727 4843 server.go:79] "Starting device plugin registration server" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.416395 4843 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.416423 4843 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.416611 4843 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.416748 4843 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.416765 4843 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.428337 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.438635 4843 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.438830 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.442113 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.442166 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.442179 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.442399 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.442869 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.442946 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.443746 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.443771 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.443781 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.443876 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.444022 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.444065 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.444719 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.444769 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.444785 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446636 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446666 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446636 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446700 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446676 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446715 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.446972 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.447151 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.447211 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448235 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448262 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448296 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448350 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448373 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448387 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448514 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448656 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.448683 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.452856 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.452954 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.452960 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.453026 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.453040 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.452980 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.453652 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.453707 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.457138 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.457192 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.457212 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.468602 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="400ms" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.488915 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.488961 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.488981 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489000 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489028 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489047 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489178 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489327 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489416 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489453 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489520 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489574 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489602 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489627 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.489659 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.516740 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.518722 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.518807 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.518836 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.518892 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.519995 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591545 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591654 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591716 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591768 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591818 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591864 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591913 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.591965 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592021 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592069 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592120 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592172 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592238 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592329 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592380 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592553 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592558 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592605 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592657 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592640 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592560 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592675 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592683 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592727 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592796 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592806 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592808 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592805 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592849 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.592925 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.720825 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.722762 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.722836 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.722855 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.722902 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.723699 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.786729 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.812573 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.839016 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-233afbec8fa731dcf85fdd29de4b7992d760ba6ec13658e82c58b6c91e572ffe WatchSource:0}: Error finding container 233afbec8fa731dcf85fdd29de4b7992d760ba6ec13658e82c58b6c91e572ffe: Status 404 returned error can't find the container with id 233afbec8fa731dcf85fdd29de4b7992d760ba6ec13658e82c58b6c91e572ffe Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.839347 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.856989 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4643c8f1347a96970332250a28d22c47f8e6c7cc5b29b51c0c6c3518813def69 WatchSource:0}: Error finding container 4643c8f1347a96970332250a28d22c47f8e6c7cc5b29b51c0c6c3518813def69: Status 404 returned error can't find the container with id 4643c8f1347a96970332250a28d22c47f8e6c7cc5b29b51c0c6c3518813def69 Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.858261 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.860356 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-8245f0b186a84e862b8a6883f9583ddfd295f276f791eaa938151668ab574471 WatchSource:0}: Error finding container 8245f0b186a84e862b8a6883f9583ddfd295f276f791eaa938151668ab574471: Status 404 returned error can't find the container with id 8245f0b186a84e862b8a6883f9583ddfd295f276f791eaa938151668ab574471 Mar 14 09:10:33 crc kubenswrapper[4843]: I0314 09:10:33.864554 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:33 crc kubenswrapper[4843]: E0314 09:10:33.869812 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="800ms" Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.881471 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a154f3c9d178255c7d1d8d140c583ad73733474743649f10293115c25f3e1095 WatchSource:0}: Error finding container a154f3c9d178255c7d1d8d140c583ad73733474743649f10293115c25f3e1095: Status 404 returned error can't find the container with id a154f3c9d178255c7d1d8d140c583ad73733474743649f10293115c25f3e1095 Mar 14 09:10:33 crc kubenswrapper[4843]: W0314 09:10:33.899669 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-7e332fc3f42cffe55c1c2513c82c27e5749137740f4dfacceca2e63fe95af424 WatchSource:0}: Error finding container 7e332fc3f42cffe55c1c2513c82c27e5749137740f4dfacceca2e63fe95af424: Status 404 returned error can't find the container with id 7e332fc3f42cffe55c1c2513c82c27e5749137740f4dfacceca2e63fe95af424 Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.124900 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.126456 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.126508 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.126522 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.126563 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.127180 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.260527 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:34 crc kubenswrapper[4843]: W0314 09:10:34.262019 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.262147 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.344758 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4643c8f1347a96970332250a28d22c47f8e6c7cc5b29b51c0c6c3518813def69"} Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.348912 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"233afbec8fa731dcf85fdd29de4b7992d760ba6ec13658e82c58b6c91e572ffe"} Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.350878 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e332fc3f42cffe55c1c2513c82c27e5749137740f4dfacceca2e63fe95af424"} Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.352577 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a154f3c9d178255c7d1d8d140c583ad73733474743649f10293115c25f3e1095"} Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.353717 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8245f0b186a84e862b8a6883f9583ddfd295f276f791eaa938151668ab574471"} Mar 14 09:10:34 crc kubenswrapper[4843]: W0314 09:10:34.575601 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.575719 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:34 crc kubenswrapper[4843]: W0314 09:10:34.609750 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.609862 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.671358 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="1.6s" Mar 14 09:10:34 crc kubenswrapper[4843]: W0314 09:10:34.763850 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.763982 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.927808 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.930371 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.930431 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.930451 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:34 crc kubenswrapper[4843]: I0314 09:10:34.930489 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:34 crc kubenswrapper[4843]: E0314 09:10:34.931174 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.199334 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 14 09:10:35 crc kubenswrapper[4843]: E0314 09:10:35.200727 4843 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.260561 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.359798 4843 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="565636033f5cab4ff2bb0fe47b0bc98bccc909fddddf436b469d9899a2dab920" exitCode=0 Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.359903 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"565636033f5cab4ff2bb0fe47b0bc98bccc909fddddf436b469d9899a2dab920"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.360015 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.361634 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.361684 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.361702 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.362742 4843 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3e323ac14cb1738c23a820eddfa32d9b88176e66c1d6c04f6556567cbb02b6a4" exitCode=0 Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.362792 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3e323ac14cb1738c23a820eddfa32d9b88176e66c1d6c04f6556567cbb02b6a4"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.362833 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.363720 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.363757 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.363772 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.366946 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a60a14f666c008ee8d8be117f482f702e2803ebfa9b49343e07f7520dccf1969"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.367010 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"55ecf455a5a502552ea20505611c9b9df9676f49afce0b88855853ff356c516d"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.367026 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c2c83d7114544865d6fe66909ce703a28d8428bde328863ec8071bbb2e53adec"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.370136 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="025e1eececd2799bfda2d8956afccae562ed75627c225a98db4a483e76433fbd" exitCode=0 Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.370199 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"025e1eececd2799bfda2d8956afccae562ed75627c225a98db4a483e76433fbd"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.370243 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.371132 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.371177 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.371191 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.372653 4843 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4aa7d21496d2c3d749be48a214bd022e197446687f094ff22c56282123714e8f" exitCode=0 Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.372713 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4aa7d21496d2c3d749be48a214bd022e197446687f094ff22c56282123714e8f"} Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.373065 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.373723 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.374404 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.374439 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.374453 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.374956 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.374984 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:35 crc kubenswrapper[4843]: I0314 09:10:35.374995 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.260662 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:36 crc kubenswrapper[4843]: E0314 09:10:36.271987 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="3.2s" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.377245 4843 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fff39fcbf93fb19840456f5599a41ead62896c9ad213252510128dd98c888500" exitCode=0 Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.377318 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fff39fcbf93fb19840456f5599a41ead62896c9ad213252510128dd98c888500"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.377433 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.378334 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.378373 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.378386 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.380808 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c6750e8842f3e3ca6a5a6405442483e8394268206eeac376351cff4c5bd606ef"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.380839 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.382101 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.382143 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.382159 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.385003 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"65fcb84a2dee472fc2f720c40429b1fa66ea919de0a815ed03f0bf68e67ec829"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.385042 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3355a240135d19759c919de0e32f623e3fa4c295606a81c1475404631f59ce2f"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.385063 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1f45d9811e4ec6ad2bb6eaf0fea0214b5d5601a30c32fcaabe2eb7563c57cf6c"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.385200 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:36 crc kubenswrapper[4843]: W0314 09:10:36.389134 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:36 crc kubenswrapper[4843]: E0314 09:10:36.389262 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.389359 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.389432 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.389461 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.392644 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d7d061f10e922b845292006b9194aaa9aff533bc7625da175f51dd113643c62"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.392797 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.396840 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.396890 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.396909 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.402234 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a4aa9fcdb00dbc4dcb3869d76cbfdfede25f06447edefc02107913ab84faf02"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.402267 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"da2bed11209848aa1fce423d8809128a64e548ac1395de019d1ca614d4973423"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.402284 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9fd4021be66e4d56e2f5bfc26c24aacc52eaa4eb7c38fdda57c52422f7d9fd68"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.402301 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"19b57e62b0b1cbb6a7798bc999bfe302198c071e3447232baebdb5e2d8af69e6"} Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.531329 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.532807 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.532863 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.532881 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:36 crc kubenswrapper[4843]: I0314 09:10:36.532976 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:36 crc kubenswrapper[4843]: E0314 09:10:36.533660 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Mar 14 09:10:36 crc kubenswrapper[4843]: W0314 09:10:36.700015 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Mar 14 09:10:36 crc kubenswrapper[4843]: E0314 09:10:36.700160 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.416009 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c2a5bd8cfc5488a8dc155299f0b25f306a1bff6d1f78affbaa1f670b46369659"} Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.416109 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.417461 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.417535 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.417559 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420349 4843 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3f4ff1fc3f598aa422b05b954355aa76192d5c172082ac193607212a6b6264bb" exitCode=0 Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420461 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3f4ff1fc3f598aa422b05b954355aa76192d5c172082ac193607212a6b6264bb"} Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420522 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420543 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420593 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420629 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.420547 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.422221 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.422280 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.422299 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.422517 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.422586 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.422654 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.423269 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.423347 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.423360 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.423409 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.423383 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:37 crc kubenswrapper[4843]: I0314 09:10:37.423489 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.305651 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.426258 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e8a5bf41694a92ec314e27587629540e9114def12df1f2ca45b7b94ae052fb9a"} Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.426346 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f1693a488671df1f228f34dcf86564663c789a767ce1cff7a05a644b45b3fd65"} Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.426363 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2ac73a8857b1fc9d1b5f23553a1b8f5d2f9a1c0e4d695a904f4294edcfadb543"} Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.426374 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.426445 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.427864 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.427939 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:38 crc kubenswrapper[4843]: I0314 09:10:38.427965 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.439701 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.439793 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.439845 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.439687 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"82b711a6525bb377faa8e5a53b106d63ef12195a13570b2fb771c9a1e15ca8c0"} Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.440048 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"56e498fa15c46a45e71751965e13d0185533b2d570f8846942c01f9847015d51"} Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.441322 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.441362 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.441375 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.441447 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.441504 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.441523 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.584953 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.733800 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.735817 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.735890 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.735906 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.735949 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.775036 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.775276 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.776709 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.776740 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:39 crc kubenswrapper[4843]: I0314 09:10:39.776751 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.048488 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.257324 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.257612 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.259482 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.259543 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.259570 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.298025 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.442036 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.442136 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.443682 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.443732 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.443753 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.443843 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.443880 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:40 crc kubenswrapper[4843]: I0314 09:10:40.443914 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.387221 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.387511 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.389430 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.389531 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.389558 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.445017 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.446390 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.446469 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:41 crc kubenswrapper[4843]: I0314 09:10:41.446495 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:42 crc kubenswrapper[4843]: I0314 09:10:42.247212 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:42 crc kubenswrapper[4843]: I0314 09:10:42.247513 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:42 crc kubenswrapper[4843]: I0314 09:10:42.249342 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:42 crc kubenswrapper[4843]: I0314 09:10:42.249402 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:42 crc kubenswrapper[4843]: I0314 09:10:42.249420 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:43 crc kubenswrapper[4843]: E0314 09:10:43.429390 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:10:44 crc kubenswrapper[4843]: I0314 09:10:44.388412 4843 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:10:44 crc kubenswrapper[4843]: I0314 09:10:44.388521 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.146151 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.146457 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.148252 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.148364 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.148433 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.155377 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.455714 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.457024 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.457064 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.457072 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:45 crc kubenswrapper[4843]: I0314 09:10:45.462148 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:46 crc kubenswrapper[4843]: I0314 09:10:46.068315 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:10:46 crc kubenswrapper[4843]: I0314 09:10:46.458790 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:46 crc kubenswrapper[4843]: I0314 09:10:46.461017 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:46 crc kubenswrapper[4843]: I0314 09:10:46.461081 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:46 crc kubenswrapper[4843]: I0314 09:10:46.461100 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:46 crc kubenswrapper[4843]: W0314 09:10:46.880972 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 14 09:10:46 crc kubenswrapper[4843]: I0314 09:10:46.881089 4843 trace.go:236] Trace[1598590347]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Mar-2026 09:10:36.879) (total time: 10001ms): Mar 14 09:10:46 crc kubenswrapper[4843]: Trace[1598590347]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:10:46.880) Mar 14 09:10:46 crc kubenswrapper[4843]: Trace[1598590347]: [10.001935818s] [10.001935818s] END Mar 14 09:10:46 crc kubenswrapper[4843]: E0314 09:10:46.881117 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.260508 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 14 09:10:47 crc kubenswrapper[4843]: W0314 09:10:47.441983 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.442092 4843 trace.go:236] Trace[458220904]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Mar-2026 09:10:37.440) (total time: 10001ms): Mar 14 09:10:47 crc kubenswrapper[4843]: Trace[458220904]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:10:47.441) Mar 14 09:10:47 crc kubenswrapper[4843]: Trace[458220904]: [10.001682493s] [10.001682493s] END Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.442141 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.461318 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.462590 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.462631 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.462640 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.819003 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:10:47 crc kubenswrapper[4843]: W0314 09:10:47.820674 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.820789 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.822293 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:10:47 crc kubenswrapper[4843]: W0314 09:10:47.827470 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.827566 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.831709 4843 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.831824 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.836862 4843 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 14 09:10:47 crc kubenswrapper[4843]: I0314 09:10:47.836938 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.839841 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 14 09:10:47 crc kubenswrapper[4843]: E0314 09:10:47.843781 4843 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.265613 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:48Z is after 2026-02-23T05:33:13Z Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.363338 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.363891 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.365744 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.366224 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.366238 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.407033 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.466583 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.469353 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c2a5bd8cfc5488a8dc155299f0b25f306a1bff6d1f78affbaa1f670b46369659" exitCode=255 Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.469529 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.470500 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c2a5bd8cfc5488a8dc155299f0b25f306a1bff6d1f78affbaa1f670b46369659"} Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.470659 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.471145 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.471178 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.471191 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.471618 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.471774 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.471850 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.472787 4843 scope.go:117] "RemoveContainer" containerID="c2a5bd8cfc5488a8dc155299f0b25f306a1bff6d1f78affbaa1f670b46369659" Mar 14 09:10:48 crc kubenswrapper[4843]: I0314 09:10:48.515707 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.248182 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.264680 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:49Z is after 2026-02-23T05:33:13Z Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.476462 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.479851 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.480490 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43"} Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.480568 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.481450 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.481515 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.481527 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.486876 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.486953 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:49 crc kubenswrapper[4843]: I0314 09:10:49.486981 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.264080 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:50Z is after 2026-02-23T05:33:13Z Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.298882 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:50 crc kubenswrapper[4843]: W0314 09:10:50.345574 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:50Z is after 2026-02-23T05:33:13Z Mar 14 09:10:50 crc kubenswrapper[4843]: E0314 09:10:50.345693 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:50Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.485439 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.486497 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.489519 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" exitCode=255 Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.489604 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43"} Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.489678 4843 scope.go:117] "RemoveContainer" containerID="c2a5bd8cfc5488a8dc155299f0b25f306a1bff6d1f78affbaa1f670b46369659" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.489717 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.491220 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.491301 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.491324 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:50 crc kubenswrapper[4843]: I0314 09:10:50.492321 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:10:50 crc kubenswrapper[4843]: E0314 09:10:50.492653 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.266023 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:51Z is after 2026-02-23T05:33:13Z Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.494904 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.497396 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.498593 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.498658 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.498684 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:51 crc kubenswrapper[4843]: I0314 09:10:51.499722 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:10:51 crc kubenswrapper[4843]: E0314 09:10:51.500023 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:10:51 crc kubenswrapper[4843]: W0314 09:10:51.670253 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:51Z is after 2026-02-23T05:33:13Z Mar 14 09:10:51 crc kubenswrapper[4843]: E0314 09:10:51.670421 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:51Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.256839 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.265535 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:52Z is after 2026-02-23T05:33:13Z Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.501780 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.503448 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.503499 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.503510 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.504158 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:10:52 crc kubenswrapper[4843]: E0314 09:10:52.504377 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:10:52 crc kubenswrapper[4843]: I0314 09:10:52.507388 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:53 crc kubenswrapper[4843]: I0314 09:10:53.264956 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:53Z is after 2026-02-23T05:33:13Z Mar 14 09:10:53 crc kubenswrapper[4843]: E0314 09:10:53.429592 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:10:53 crc kubenswrapper[4843]: I0314 09:10:53.504865 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:53 crc kubenswrapper[4843]: I0314 09:10:53.506162 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:53 crc kubenswrapper[4843]: I0314 09:10:53.506225 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:53 crc kubenswrapper[4843]: I0314 09:10:53.506243 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:53 crc kubenswrapper[4843]: I0314 09:10:53.507220 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:10:53 crc kubenswrapper[4843]: E0314 09:10:53.507523 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.220554 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.222363 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.222414 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.222435 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.222504 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:10:54 crc kubenswrapper[4843]: E0314 09:10:54.227878 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:54Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:10:54 crc kubenswrapper[4843]: E0314 09:10:54.246010 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:54Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.264947 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:54Z is after 2026-02-23T05:33:13Z Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.388549 4843 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:10:54 crc kubenswrapper[4843]: I0314 09:10:54.388657 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:10:55 crc kubenswrapper[4843]: I0314 09:10:55.263829 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:55Z is after 2026-02-23T05:33:13Z Mar 14 09:10:56 crc kubenswrapper[4843]: I0314 09:10:56.114776 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 14 09:10:56 crc kubenswrapper[4843]: E0314 09:10:56.120195 4843 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:56 crc kubenswrapper[4843]: I0314 09:10:56.266001 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:56Z is after 2026-02-23T05:33:13Z Mar 14 09:10:57 crc kubenswrapper[4843]: I0314 09:10:57.264521 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:57Z is after 2026-02-23T05:33:13Z Mar 14 09:10:57 crc kubenswrapper[4843]: E0314 09:10:57.829780 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:57Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:10:58 crc kubenswrapper[4843]: I0314 09:10:58.267030 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:58Z is after 2026-02-23T05:33:13Z Mar 14 09:10:58 crc kubenswrapper[4843]: W0314 09:10:58.588927 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:58Z is after 2026-02-23T05:33:13Z Mar 14 09:10:58 crc kubenswrapper[4843]: E0314 09:10:58.589085 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:58Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:59 crc kubenswrapper[4843]: W0314 09:10:59.194127 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:59Z is after 2026-02-23T05:33:13Z Mar 14 09:10:59 crc kubenswrapper[4843]: E0314 09:10:59.194212 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.247766 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.248020 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.249709 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.249782 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.249808 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.250904 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:10:59 crc kubenswrapper[4843]: E0314 09:10:59.251266 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:10:59 crc kubenswrapper[4843]: I0314 09:10:59.265874 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:59Z is after 2026-02-23T05:33:13Z Mar 14 09:10:59 crc kubenswrapper[4843]: W0314 09:10:59.394073 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:59Z is after 2026-02-23T05:33:13Z Mar 14 09:10:59 crc kubenswrapper[4843]: E0314 09:10:59.394210 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:10:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:00 crc kubenswrapper[4843]: I0314 09:11:00.265428 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:00Z is after 2026-02-23T05:33:13Z Mar 14 09:11:00 crc kubenswrapper[4843]: W0314 09:11:00.499331 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:00Z is after 2026-02-23T05:33:13Z Mar 14 09:11:00 crc kubenswrapper[4843]: E0314 09:11:00.499439 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:01 crc kubenswrapper[4843]: I0314 09:11:01.228390 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:01 crc kubenswrapper[4843]: I0314 09:11:01.230495 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:01 crc kubenswrapper[4843]: I0314 09:11:01.230728 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:01 crc kubenswrapper[4843]: I0314 09:11:01.230881 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:01 crc kubenswrapper[4843]: I0314 09:11:01.231040 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:01 crc kubenswrapper[4843]: E0314 09:11:01.236352 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:01Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:11:01 crc kubenswrapper[4843]: E0314 09:11:01.252019 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:01Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 14 09:11:01 crc kubenswrapper[4843]: I0314 09:11:01.265392 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:01Z is after 2026-02-23T05:33:13Z Mar 14 09:11:02 crc kubenswrapper[4843]: I0314 09:11:02.265316 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:02Z is after 2026-02-23T05:33:13Z Mar 14 09:11:03 crc kubenswrapper[4843]: I0314 09:11:03.265330 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:03Z is after 2026-02-23T05:33:13Z Mar 14 09:11:03 crc kubenswrapper[4843]: E0314 09:11:03.430155 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.265253 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:04Z is after 2026-02-23T05:33:13Z Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.388469 4843 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.388630 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.388767 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.389030 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.390577 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.390635 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.390657 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.391494 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"55ecf455a5a502552ea20505611c9b9df9676f49afce0b88855853ff356c516d"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.391736 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://55ecf455a5a502552ea20505611c9b9df9676f49afce0b88855853ff356c516d" gracePeriod=30 Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.538115 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.538834 4843 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="55ecf455a5a502552ea20505611c9b9df9676f49afce0b88855853ff356c516d" exitCode=255 Mar 14 09:11:04 crc kubenswrapper[4843]: I0314 09:11:04.538894 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"55ecf455a5a502552ea20505611c9b9df9676f49afce0b88855853ff356c516d"} Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.262291 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:05Z is after 2026-02-23T05:33:13Z Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.545342 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.545866 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"de74811fdce1b447614f552a00f447fecfcf8f16728485f0f8feb89cd4b899e0"} Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.546058 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.547469 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.547516 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:05 crc kubenswrapper[4843]: I0314 09:11:05.547534 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:06 crc kubenswrapper[4843]: I0314 09:11:06.067567 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:06 crc kubenswrapper[4843]: I0314 09:11:06.265457 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:06Z is after 2026-02-23T05:33:13Z Mar 14 09:11:06 crc kubenswrapper[4843]: I0314 09:11:06.547610 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:06 crc kubenswrapper[4843]: I0314 09:11:06.548696 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:06 crc kubenswrapper[4843]: I0314 09:11:06.548742 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:06 crc kubenswrapper[4843]: I0314 09:11:06.548755 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:07 crc kubenswrapper[4843]: I0314 09:11:07.266601 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:07Z is after 2026-02-23T05:33:13Z Mar 14 09:11:07 crc kubenswrapper[4843]: I0314 09:11:07.550088 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:07 crc kubenswrapper[4843]: I0314 09:11:07.551572 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:07 crc kubenswrapper[4843]: I0314 09:11:07.551638 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:07 crc kubenswrapper[4843]: I0314 09:11:07.551650 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:07 crc kubenswrapper[4843]: E0314 09:11:07.834965 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:07Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:08 crc kubenswrapper[4843]: I0314 09:11:08.237243 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:08 crc kubenswrapper[4843]: I0314 09:11:08.239368 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:08 crc kubenswrapper[4843]: I0314 09:11:08.239566 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:08 crc kubenswrapper[4843]: I0314 09:11:08.239594 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:08 crc kubenswrapper[4843]: I0314 09:11:08.239637 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:08 crc kubenswrapper[4843]: E0314 09:11:08.244911 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:08Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:11:08 crc kubenswrapper[4843]: E0314 09:11:08.258251 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:08Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 14 09:11:08 crc kubenswrapper[4843]: I0314 09:11:08.265244 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:08Z is after 2026-02-23T05:33:13Z Mar 14 09:11:09 crc kubenswrapper[4843]: I0314 09:11:09.265032 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:09Z is after 2026-02-23T05:33:13Z Mar 14 09:11:10 crc kubenswrapper[4843]: I0314 09:11:10.265028 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:10Z is after 2026-02-23T05:33:13Z Mar 14 09:11:11 crc kubenswrapper[4843]: I0314 09:11:11.265985 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:11Z is after 2026-02-23T05:33:13Z Mar 14 09:11:11 crc kubenswrapper[4843]: I0314 09:11:11.387876 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:11 crc kubenswrapper[4843]: I0314 09:11:11.388149 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:11 crc kubenswrapper[4843]: I0314 09:11:11.389857 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:11 crc kubenswrapper[4843]: I0314 09:11:11.389920 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:11 crc kubenswrapper[4843]: I0314 09:11:11.389934 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:12 crc kubenswrapper[4843]: I0314 09:11:12.265005 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:12Z is after 2026-02-23T05:33:13Z Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.265311 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:13Z is after 2026-02-23T05:33:13Z Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.338728 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.340407 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.340447 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.340456 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.341056 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.367720 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 14 09:11:13 crc kubenswrapper[4843]: E0314 09:11:13.375367 4843 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:13Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:13 crc kubenswrapper[4843]: E0314 09:11:13.376585 4843 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 14 09:11:13 crc kubenswrapper[4843]: E0314 09:11:13.430357 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.568125 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.571192 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543"} Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.571392 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.572614 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.572669 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:13 crc kubenswrapper[4843]: I0314 09:11:13.572682 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:14 crc kubenswrapper[4843]: I0314 09:11:14.263776 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:14Z is after 2026-02-23T05:33:13Z Mar 14 09:11:14 crc kubenswrapper[4843]: I0314 09:11:14.388970 4843 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:11:14 crc kubenswrapper[4843]: I0314 09:11:14.389049 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.246016 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.248034 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.248117 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.248133 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.248173 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:15 crc kubenswrapper[4843]: E0314 09:11:15.252148 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:15Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:11:15 crc kubenswrapper[4843]: E0314 09:11:15.265199 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:15Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.267832 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:15Z is after 2026-02-23T05:33:13Z Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.578492 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.579126 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.581419 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" exitCode=255 Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.581484 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543"} Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.581546 4843 scope.go:117] "RemoveContainer" containerID="f0e938b9b9e92c10e78d577e74b71d1c4305bb1c749efb6933c08a741cce9c43" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.581733 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.582938 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.582983 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.582995 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:15 crc kubenswrapper[4843]: I0314 09:11:15.583640 4843 scope.go:117] "RemoveContainer" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" Mar 14 09:11:15 crc kubenswrapper[4843]: E0314 09:11:15.583796 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:16 crc kubenswrapper[4843]: I0314 09:11:16.264323 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:16Z is after 2026-02-23T05:33:13Z Mar 14 09:11:16 crc kubenswrapper[4843]: I0314 09:11:16.587989 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 14 09:11:17 crc kubenswrapper[4843]: I0314 09:11:17.264582 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:17Z is after 2026-02-23T05:33:13Z Mar 14 09:11:17 crc kubenswrapper[4843]: E0314 09:11:17.839972 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:17Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:18 crc kubenswrapper[4843]: I0314 09:11:18.264983 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:18Z is after 2026-02-23T05:33:13Z Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.248208 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.248506 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.250060 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.250129 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.250157 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.251170 4843 scope.go:117] "RemoveContainer" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" Mar 14 09:11:19 crc kubenswrapper[4843]: E0314 09:11:19.251590 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:19 crc kubenswrapper[4843]: I0314 09:11:19.265239 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:19Z is after 2026-02-23T05:33:13Z Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.264953 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.265167 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.266415 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:20Z is after 2026-02-23T05:33:13Z Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.271428 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.271499 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.271512 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.298256 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.298531 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.299706 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.299923 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.300081 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:20 crc kubenswrapper[4843]: I0314 09:11:20.301075 4843 scope.go:117] "RemoveContainer" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" Mar 14 09:11:20 crc kubenswrapper[4843]: E0314 09:11:20.301585 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:21 crc kubenswrapper[4843]: W0314 09:11:21.180103 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:21Z is after 2026-02-23T05:33:13Z Mar 14 09:11:21 crc kubenswrapper[4843]: E0314 09:11:21.180227 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:21Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:21 crc kubenswrapper[4843]: I0314 09:11:21.266478 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:21Z is after 2026-02-23T05:33:13Z Mar 14 09:11:22 crc kubenswrapper[4843]: I0314 09:11:22.253354 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:22 crc kubenswrapper[4843]: I0314 09:11:22.255482 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:22 crc kubenswrapper[4843]: I0314 09:11:22.255562 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:22 crc kubenswrapper[4843]: I0314 09:11:22.255603 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:22 crc kubenswrapper[4843]: I0314 09:11:22.255655 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:22 crc kubenswrapper[4843]: E0314 09:11:22.260898 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:22Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:11:22 crc kubenswrapper[4843]: I0314 09:11:22.265228 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:22Z is after 2026-02-23T05:33:13Z Mar 14 09:11:22 crc kubenswrapper[4843]: E0314 09:11:22.270532 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:22Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 14 09:11:22 crc kubenswrapper[4843]: W0314 09:11:22.917733 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:22Z is after 2026-02-23T05:33:13Z Mar 14 09:11:22 crc kubenswrapper[4843]: E0314 09:11:22.917832 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:23 crc kubenswrapper[4843]: I0314 09:11:23.264232 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:23Z is after 2026-02-23T05:33:13Z Mar 14 09:11:23 crc kubenswrapper[4843]: E0314 09:11:23.430557 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:11:24 crc kubenswrapper[4843]: W0314 09:11:24.200830 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:24Z is after 2026-02-23T05:33:13Z Mar 14 09:11:24 crc kubenswrapper[4843]: E0314 09:11:24.200919 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:24 crc kubenswrapper[4843]: I0314 09:11:24.263498 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:24Z is after 2026-02-23T05:33:13Z Mar 14 09:11:24 crc kubenswrapper[4843]: I0314 09:11:24.388306 4843 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:11:24 crc kubenswrapper[4843]: I0314 09:11:24.388395 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:11:24 crc kubenswrapper[4843]: W0314 09:11:24.750758 4843 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:24Z is after 2026-02-23T05:33:13Z Mar 14 09:11:24 crc kubenswrapper[4843]: E0314 09:11:24.750872 4843 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 14 09:11:25 crc kubenswrapper[4843]: I0314 09:11:25.264253 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:25Z is after 2026-02-23T05:33:13Z Mar 14 09:11:26 crc kubenswrapper[4843]: I0314 09:11:26.264127 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:26Z is after 2026-02-23T05:33:13Z Mar 14 09:11:27 crc kubenswrapper[4843]: I0314 09:11:27.263412 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:27Z is after 2026-02-23T05:33:13Z Mar 14 09:11:27 crc kubenswrapper[4843]: E0314 09:11:27.846068 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:27Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:28 crc kubenswrapper[4843]: I0314 09:11:28.264839 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:28Z is after 2026-02-23T05:33:13Z Mar 14 09:11:29 crc kubenswrapper[4843]: I0314 09:11:29.261067 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:29 crc kubenswrapper[4843]: I0314 09:11:29.262943 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:29 crc kubenswrapper[4843]: I0314 09:11:29.263021 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:29 crc kubenswrapper[4843]: I0314 09:11:29.263037 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:29 crc kubenswrapper[4843]: I0314 09:11:29.263061 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:29 crc kubenswrapper[4843]: I0314 09:11:29.265668 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:29Z is after 2026-02-23T05:33:13Z Mar 14 09:11:29 crc kubenswrapper[4843]: E0314 09:11:29.269531 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:29Z is after 2026-02-23T05:33:13Z" node="crc" Mar 14 09:11:29 crc kubenswrapper[4843]: E0314 09:11:29.276423 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:29Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 14 09:11:30 crc kubenswrapper[4843]: I0314 09:11:30.263742 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:30Z is after 2026-02-23T05:33:13Z Mar 14 09:11:31 crc kubenswrapper[4843]: I0314 09:11:31.263207 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:31Z is after 2026-02-23T05:33:13Z Mar 14 09:11:31 crc kubenswrapper[4843]: I0314 09:11:31.338505 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:31 crc kubenswrapper[4843]: I0314 09:11:31.339894 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:31 crc kubenswrapper[4843]: I0314 09:11:31.339978 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:31 crc kubenswrapper[4843]: I0314 09:11:31.339996 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:31 crc kubenswrapper[4843]: I0314 09:11:31.340717 4843 scope.go:117] "RemoveContainer" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" Mar 14 09:11:31 crc kubenswrapper[4843]: E0314 09:11:31.340939 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:32 crc kubenswrapper[4843]: I0314 09:11:32.263292 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:32Z is after 2026-02-23T05:33:13Z Mar 14 09:11:33 crc kubenswrapper[4843]: I0314 09:11:33.262335 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:33Z is after 2026-02-23T05:33:13Z Mar 14 09:11:33 crc kubenswrapper[4843]: E0314 09:11:33.430665 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.263034 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:11:34Z is after 2026-02-23T05:33:13Z Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.387849 4843 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.387987 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.388082 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.388342 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.389921 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.389979 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.389990 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.390666 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"de74811fdce1b447614f552a00f447fecfcf8f16728485f0f8feb89cd4b899e0"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.390772 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://de74811fdce1b447614f552a00f447fecfcf8f16728485f0f8feb89cd4b899e0" gracePeriod=30 Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.641849 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.643621 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.644041 4843 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="de74811fdce1b447614f552a00f447fecfcf8f16728485f0f8feb89cd4b899e0" exitCode=255 Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.644077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"de74811fdce1b447614f552a00f447fecfcf8f16728485f0f8feb89cd4b899e0"} Mar 14 09:11:34 crc kubenswrapper[4843]: I0314 09:11:34.644134 4843 scope.go:117] "RemoveContainer" containerID="55ecf455a5a502552ea20505611c9b9df9676f49afce0b88855853ff356c516d" Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.267055 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.655331 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.656687 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"76d0999b6b70ea5c90638ba52bbec6e5bbe0e904f3081bfefbf3de6bd819bf7c"} Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.656904 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.658201 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.658318 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:35 crc kubenswrapper[4843]: I0314 09:11:35.658397 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.067957 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.266083 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.269888 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.271834 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.271895 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.271914 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.271952 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:36 crc kubenswrapper[4843]: E0314 09:11:36.279966 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 14 09:11:36 crc kubenswrapper[4843]: E0314 09:11:36.280052 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.659821 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.661410 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.661493 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:36 crc kubenswrapper[4843]: I0314 09:11:36.661513 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:37 crc kubenswrapper[4843]: I0314 09:11:37.266688 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:37 crc kubenswrapper[4843]: I0314 09:11:37.662035 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:37 crc kubenswrapper[4843]: I0314 09:11:37.662993 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:37 crc kubenswrapper[4843]: I0314 09:11:37.663050 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:37 crc kubenswrapper[4843]: I0314 09:11:37.663066 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.851701 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f3c3ad5f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,LastTimestamp:2026-03-14 09:10:33.257332063 +0000 UTC m=+0.569943261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.856957 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.862203 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.867108 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.871777 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25fe3d8583 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.433089411 +0000 UTC m=+0.745700539,LastTimestamp:2026-03-14 09:10:33.433089411 +0000 UTC m=+0.745700539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.877232 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.442150027 +0000 UTC m=+0.754761175,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.884211 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.442174338 +0000 UTC m=+0.754785476,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.888888 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7490066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.442186408 +0000 UTC m=+0.754797556,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.896999 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.443763587 +0000 UTC m=+0.756374715,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.902065 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.443777118 +0000 UTC m=+0.756388246,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.908338 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7490066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.443786838 +0000 UTC m=+0.756397966,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.913017 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.444756792 +0000 UTC m=+0.757367940,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.917900 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.444778903 +0000 UTC m=+0.757390041,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.925664 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7490066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.444792313 +0000 UTC m=+0.757403451,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.931778 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.446658559 +0000 UTC m=+0.759269687,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.939134 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.4466725 +0000 UTC m=+0.759283628,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.944681 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.44668796 +0000 UTC m=+0.759299098,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.950234 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.446709571 +0000 UTC m=+0.759320709,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.955498 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7490066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.446746381 +0000 UTC m=+0.759357559,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.962455 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7490066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.446776712 +0000 UTC m=+0.759387850,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.968118 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.448254369 +0000 UTC m=+0.760865497,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.974566 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.44829147 +0000 UTC m=+0.760902588,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.980000 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7490066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7490066 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316401254 +0000 UTC m=+0.629012382,LastTimestamp:2026-03-14 09:10:33.44830142 +0000 UTC m=+0.760912548,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.986785 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f7482e0b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f7482e0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316347403 +0000 UTC m=+0.628958531,LastTimestamp:2026-03-14 09:10:33.448365542 +0000 UTC m=+0.760976680,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:37 crc kubenswrapper[4843]: E0314 09:11:37.993950 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189caa25f748d370\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189caa25f748d370 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.316389744 +0000 UTC m=+0.629000872,LastTimestamp:2026-03-14 09:10:33.448381062 +0000 UTC m=+0.760992200,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.001184 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa26171c37a8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.850337192 +0000 UTC m=+1.162948350,LastTimestamp:2026-03-14 09:10:33.850337192 +0000 UTC m=+1.162948350,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.006435 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa2617ef7d96 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.86418319 +0000 UTC m=+1.176794328,LastTimestamp:2026-03-14 09:10:33.86418319 +0000 UTC m=+1.176794328,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.011647 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189caa2617f0b76c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.864263532 +0000 UTC m=+1.176874700,LastTimestamp:2026-03-14 09:10:33.864263532 +0000 UTC m=+1.176874700,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.017493 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa261954a3f8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.887589368 +0000 UTC m=+1.200200506,LastTimestamp:2026-03-14 09:10:33.887589368 +0000 UTC m=+1.200200506,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.024232 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa261a5f03e0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:33.905046496 +0000 UTC m=+1.217657654,LastTimestamp:2026-03-14 09:10:33.905046496 +0000 UTC m=+1.217657654,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.029569 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa263e91789c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.512332956 +0000 UTC m=+1.824944094,LastTimestamp:2026-03-14 09:10:34.512332956 +0000 UTC m=+1.824944094,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.034493 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189caa263ea930b7 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.513887415 +0000 UTC m=+1.826498553,LastTimestamp:2026-03-14 09:10:34.513887415 +0000 UTC m=+1.826498553,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.039574 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa263eb0cd66 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.514386278 +0000 UTC m=+1.826997416,LastTimestamp:2026-03-14 09:10:34.514386278 +0000 UTC m=+1.826997416,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.044898 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa263eb1d852 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.51445461 +0000 UTC m=+1.827065748,LastTimestamp:2026-03-14 09:10:34.51445461 +0000 UTC m=+1.827065748,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.050580 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa263eb70fb4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.514796468 +0000 UTC m=+1.827407606,LastTimestamp:2026-03-14 09:10:34.514796468 +0000 UTC m=+1.827407606,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.054917 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa263f6ca035 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.526695477 +0000 UTC m=+1.839306625,LastTimestamp:2026-03-14 09:10:34.526695477 +0000 UTC m=+1.839306625,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.062358 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa263f7385d7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.527147479 +0000 UTC m=+1.839758617,LastTimestamp:2026-03-14 09:10:34.527147479 +0000 UTC m=+1.839758617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.067480 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa263f9a8b31 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.529704753 +0000 UTC m=+1.842315891,LastTimestamp:2026-03-14 09:10:34.529704753 +0000 UTC m=+1.842315891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.072452 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa263fd650e7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.533621991 +0000 UTC m=+1.846233129,LastTimestamp:2026-03-14 09:10:34.533621991 +0000 UTC m=+1.846233129,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.077110 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189caa263ffa782d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.535991341 +0000 UTC m=+1.848602479,LastTimestamp:2026-03-14 09:10:34.535991341 +0000 UTC m=+1.848602479,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.081333 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa263fff3857 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.536302679 +0000 UTC m=+1.848913817,LastTimestamp:2026-03-14 09:10:34.536302679 +0000 UTC m=+1.848913817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.084894 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa265623740e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.907776014 +0000 UTC m=+2.220387182,LastTimestamp:2026-03-14 09:10:34.907776014 +0000 UTC m=+2.220387182,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.089864 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2656ff72a8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.922193576 +0000 UTC m=+2.234804734,LastTimestamp:2026-03-14 09:10:34.922193576 +0000 UTC m=+2.234804734,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.096285 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa26571571ee openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.923635182 +0000 UTC m=+2.236246320,LastTimestamp:2026-03-14 09:10:34.923635182 +0000 UTC m=+2.236246320,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.100843 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2668d4c134 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.221385524 +0000 UTC m=+2.533996662,LastTimestamp:2026-03-14 09:10:35.221385524 +0000 UTC m=+2.533996662,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.105329 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2669d4dc24 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.238169636 +0000 UTC m=+2.550780764,LastTimestamp:2026-03-14 09:10:35.238169636 +0000 UTC m=+2.550780764,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.109320 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2669ed0631 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.239753265 +0000 UTC m=+2.552364403,LastTimestamp:2026-03-14 09:10:35.239753265 +0000 UTC m=+2.552364403,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.113345 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189caa26714ec993 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.363600787 +0000 UTC m=+2.676211935,LastTimestamp:2026-03-14 09:10:35.363600787 +0000 UTC m=+2.676211935,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.116671 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa26716a088d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.365386381 +0000 UTC m=+2.677997509,LastTimestamp:2026-03-14 09:10:35.365386381 +0000 UTC m=+2.677997509,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.122327 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa2671e5987b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.373484155 +0000 UTC m=+2.686095283,LastTimestamp:2026-03-14 09:10:35.373484155 +0000 UTC m=+2.686095283,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.129847 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26720ea9cb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.376175563 +0000 UTC m=+2.688786681,LastTimestamp:2026-03-14 09:10:35.376175563 +0000 UTC m=+2.688786681,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.134519 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2679344eca openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.496083146 +0000 UTC m=+2.808694274,LastTimestamp:2026-03-14 09:10:35.496083146 +0000 UTC m=+2.808694274,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.139567 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa267a8803de openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.518346206 +0000 UTC m=+2.830957334,LastTimestamp:2026-03-14 09:10:35.518346206 +0000 UTC m=+2.830957334,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.152713 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189caa267d3ba4ee openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.563672814 +0000 UTC m=+2.876283942,LastTimestamp:2026-03-14 09:10:35.563672814 +0000 UTC m=+2.876283942,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.158195 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa267d41a880 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.564066944 +0000 UTC m=+2.876678072,LastTimestamp:2026-03-14 09:10:35.564066944 +0000 UTC m=+2.876678072,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.164510 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa267d95a8de openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.569572062 +0000 UTC m=+2.882183190,LastTimestamp:2026-03-14 09:10:35.569572062 +0000 UTC m=+2.882183190,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.170767 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189caa267deeff9a openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.57542697 +0000 UTC m=+2.888038098,LastTimestamp:2026-03-14 09:10:35.57542697 +0000 UTC m=+2.888038098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.176810 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa267e09ffa6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.577196454 +0000 UTC m=+2.889807582,LastTimestamp:2026-03-14 09:10:35.577196454 +0000 UTC m=+2.889807582,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.182968 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa267e1b5d31 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.578334513 +0000 UTC m=+2.890945641,LastTimestamp:2026-03-14 09:10:35.578334513 +0000 UTC m=+2.890945641,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.186942 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa267eedb5ac openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.592119724 +0000 UTC m=+2.904730852,LastTimestamp:2026-03-14 09:10:35.592119724 +0000 UTC m=+2.904730852,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.193017 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa267f1da26d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.595260525 +0000 UTC m=+2.907871653,LastTimestamp:2026-03-14 09:10:35.595260525 +0000 UTC m=+2.907871653,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.198598 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26803f389c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.614238876 +0000 UTC m=+2.926850024,LastTimestamp:2026-03-14 09:10:35.614238876 +0000 UTC m=+2.926850024,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.204788 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa268191e1c6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.63643335 +0000 UTC m=+2.949044488,LastTimestamp:2026-03-14 09:10:35.63643335 +0000 UTC m=+2.949044488,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.209030 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa268cb0525d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.822977629 +0000 UTC m=+3.135588757,LastTimestamp:2026-03-14 09:10:35.822977629 +0000 UTC m=+3.135588757,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.214508 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa268d131275 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.829449333 +0000 UTC m=+3.142060461,LastTimestamp:2026-03-14 09:10:35.829449333 +0000 UTC m=+3.142060461,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.218670 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa268d5193b6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.833545654 +0000 UTC m=+3.146156782,LastTimestamp:2026-03-14 09:10:35.833545654 +0000 UTC m=+3.146156782,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.222332 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa268d614c9b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.834576027 +0000 UTC m=+3.147187155,LastTimestamp:2026-03-14 09:10:35.834576027 +0000 UTC m=+3.147187155,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.226480 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa268e41654d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.849262413 +0000 UTC m=+3.161873541,LastTimestamp:2026-03-14 09:10:35.849262413 +0000 UTC m=+3.161873541,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.231012 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa268e548e9b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:35.850518171 +0000 UTC m=+3.163129299,LastTimestamp:2026-03-14 09:10:35.850518171 +0000 UTC m=+3.163129299,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.235389 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa269a5987a0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.052170656 +0000 UTC m=+3.364781794,LastTimestamp:2026-03-14 09:10:36.052170656 +0000 UTC m=+3.364781794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.240103 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa269a728c24 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.053810212 +0000 UTC m=+3.366421380,LastTimestamp:2026-03-14 09:10:36.053810212 +0000 UTC m=+3.366421380,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.243651 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189caa269b0ab816 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.063782934 +0000 UTC m=+3.376394092,LastTimestamp:2026-03-14 09:10:36.063782934 +0000 UTC m=+3.376394092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.246615 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa269bd3a13a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.076949818 +0000 UTC m=+3.389560966,LastTimestamp:2026-03-14 09:10:36.076949818 +0000 UTC m=+3.389560966,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.250007 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa269be59508 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.078126344 +0000 UTC m=+3.390737492,LastTimestamp:2026-03-14 09:10:36.078126344 +0000 UTC m=+3.390737492,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.253384 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26a9540116 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.303466774 +0000 UTC m=+3.616077912,LastTimestamp:2026-03-14 09:10:36.303466774 +0000 UTC m=+3.616077912,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.257601 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26aa0bdc4f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.315515983 +0000 UTC m=+3.628127121,LastTimestamp:2026-03-14 09:10:36.315515983 +0000 UTC m=+3.628127121,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: I0314 09:11:38.262152 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.262202 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26aa1ea066 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.31674583 +0000 UTC m=+3.629356968,LastTimestamp:2026-03-14 09:10:36.31674583 +0000 UTC m=+3.629356968,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.263583 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26ade09fc6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.379791302 +0000 UTC m=+3.692402430,LastTimestamp:2026-03-14 09:10:36.379791302 +0000 UTC m=+3.692402430,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.267625 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26b6c76158 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.529131864 +0000 UTC m=+3.841743002,LastTimestamp:2026-03-14 09:10:36.529131864 +0000 UTC m=+3.841743002,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.271251 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26b795a3ce openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.542649294 +0000 UTC m=+3.855260442,LastTimestamp:2026-03-14 09:10:36.542649294 +0000 UTC m=+3.855260442,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.274807 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26bcc6f9a4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.629768612 +0000 UTC m=+3.942379740,LastTimestamp:2026-03-14 09:10:36.629768612 +0000 UTC m=+3.942379740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.278577 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26bd7193f9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.640949241 +0000 UTC m=+3.953560369,LastTimestamp:2026-03-14 09:10:36.640949241 +0000 UTC m=+3.953560369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.283643 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26ec28e3ae openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.42471467 +0000 UTC m=+4.737325838,LastTimestamp:2026-03-14 09:10:37.42471467 +0000 UTC m=+4.737325838,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.287150 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26fb1a4840 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.675415616 +0000 UTC m=+4.988026744,LastTimestamp:2026-03-14 09:10:37.675415616 +0000 UTC m=+4.988026744,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.290726 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26fbe68304 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.688800004 +0000 UTC m=+5.001411142,LastTimestamp:2026-03-14 09:10:37.688800004 +0000 UTC m=+5.001411142,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.293968 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa26fc032c0d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.690678285 +0000 UTC m=+5.003289413,LastTimestamp:2026-03-14 09:10:37.690678285 +0000 UTC m=+5.003289413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.297556 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa270a7a78a8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.933377704 +0000 UTC m=+5.245988872,LastTimestamp:2026-03-14 09:10:37.933377704 +0000 UTC m=+5.245988872,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.301373 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa270bb395a1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.953897889 +0000 UTC m=+5.266509067,LastTimestamp:2026-03-14 09:10:37.953897889 +0000 UTC m=+5.266509067,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.305173 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa270bcfdbc3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:37.955750851 +0000 UTC m=+5.268361989,LastTimestamp:2026-03-14 09:10:37.955750851 +0000 UTC m=+5.268361989,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.310080 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa2719d82b6d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.191176557 +0000 UTC m=+5.503787685,LastTimestamp:2026-03-14 09:10:38.191176557 +0000 UTC m=+5.503787685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.313952 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa271a89954f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.202803535 +0000 UTC m=+5.515414663,LastTimestamp:2026-03-14 09:10:38.202803535 +0000 UTC m=+5.515414663,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.317813 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa271a9d51e2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.204096994 +0000 UTC m=+5.516708122,LastTimestamp:2026-03-14 09:10:38.204096994 +0000 UTC m=+5.516708122,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.322575 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa2727f22de4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.427762148 +0000 UTC m=+5.740373276,LastTimestamp:2026-03-14 09:10:38.427762148 +0000 UTC m=+5.740373276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.326676 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa2728cebf99 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.442217369 +0000 UTC m=+5.754828527,LastTimestamp:2026-03-14 09:10:38.442217369 +0000 UTC m=+5.754828527,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.330036 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa2728e8dbc5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.443928517 +0000 UTC m=+5.756539655,LastTimestamp:2026-03-14 09:10:38.443928517 +0000 UTC m=+5.756539655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.333618 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa273379c250 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.62119688 +0000 UTC m=+5.933808008,LastTimestamp:2026-03-14 09:10:38.62119688 +0000 UTC m=+5.933808008,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.337131 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189caa273436c8c4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:38.633584836 +0000 UTC m=+5.946195984,LastTimestamp:2026-03-14 09:10:38.633584836 +0000 UTC m=+5.946195984,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.346497 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-controller-manager-crc.189caa288b3b9e24 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 14 09:11:38 crc kubenswrapper[4843]: body: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:44.388486692 +0000 UTC m=+11.701097840,LastTimestamp:2026-03-14 09:10:44.388486692 +0000 UTC m=+11.701097840,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.351486 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa288b3cc5c6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:44.388562374 +0000 UTC m=+11.701173502,LastTimestamp:2026-03-14 09:10:44.388562374 +0000 UTC m=+11.701173502,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.357350 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-apiserver-crc.189caa29587854a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 14 09:11:38 crc kubenswrapper[4843]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 14 09:11:38 crc kubenswrapper[4843]: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:47.83179485 +0000 UTC m=+15.144406008,LastTimestamp:2026-03-14 09:10:47.83179485 +0000 UTC m=+15.144406008,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.361050 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa2958797757 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:47.831869271 +0000 UTC m=+15.144480429,LastTimestamp:2026-03-14 09:10:47.831869271 +0000 UTC m=+15.144480429,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.364443 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189caa29587854a2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-apiserver-crc.189caa29587854a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 14 09:11:38 crc kubenswrapper[4843]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 14 09:11:38 crc kubenswrapper[4843]: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:47.83179485 +0000 UTC m=+15.144406008,LastTimestamp:2026-03-14 09:10:47.836914764 +0000 UTC m=+15.149525912,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.369149 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189caa2958797757\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa2958797757 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:47.831869271 +0000 UTC m=+15.144480429,LastTimestamp:2026-03-14 09:10:47.836991535 +0000 UTC m=+15.149602683,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.378380 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189caa26aa1ea066\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26aa1ea066 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.31674583 +0000 UTC m=+3.629356968,LastTimestamp:2026-03-14 09:10:48.486809267 +0000 UTC m=+15.799420405,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.384507 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189caa26b6c76158\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26b6c76158 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.529131864 +0000 UTC m=+3.841743002,LastTimestamp:2026-03-14 09:10:48.690784236 +0000 UTC m=+16.003395394,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.390439 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189caa26b795a3ce\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189caa26b795a3ce openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:36.542649294 +0000 UTC m=+3.855260442,LastTimestamp:2026-03-14 09:10:48.708520514 +0000 UTC m=+16.021131672,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.397253 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf49a997 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 14 09:11:38 crc kubenswrapper[4843]: body: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388627863 +0000 UTC m=+21.701239031,LastTimestamp:2026-03-14 09:10:54.388627863 +0000 UTC m=+21.701239031,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.401748 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf4ad297 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388703895 +0000 UTC m=+21.701315063,LastTimestamp:2026-03-14 09:10:54.388703895 +0000 UTC m=+21.701315063,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.407253 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa2adf49a997\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf49a997 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 14 09:11:38 crc kubenswrapper[4843]: body: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388627863 +0000 UTC m=+21.701239031,LastTimestamp:2026-03-14 09:11:04.388596835 +0000 UTC m=+31.701208003,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.410890 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa2adf4ad297\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf4ad297 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388703895 +0000 UTC m=+21.701315063,LastTimestamp:2026-03-14 09:11:04.388676247 +0000 UTC m=+31.701287405,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.415341 4843 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2d33849a17 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:11:04.391711255 +0000 UTC m=+31.704322413,LastTimestamp:2026-03-14 09:11:04.391711255 +0000 UTC m=+31.704322413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.419083 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa263f9a8b31\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa263f9a8b31 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.529704753 +0000 UTC m=+1.842315891,LastTimestamp:2026-03-14 09:11:04.51204312 +0000 UTC m=+31.824654248,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.423207 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa265623740e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa265623740e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.907776014 +0000 UTC m=+2.220387182,LastTimestamp:2026-03-14 09:11:04.736927702 +0000 UTC m=+32.049538870,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.426877 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa2656ff72a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2656ff72a8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:34.922193576 +0000 UTC m=+2.234804734,LastTimestamp:2026-03-14 09:11:04.75074116 +0000 UTC m=+32.063352298,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.431734 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa2adf49a997\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf49a997 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 14 09:11:38 crc kubenswrapper[4843]: body: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388627863 +0000 UTC m=+21.701239031,LastTimestamp:2026-03-14 09:11:14.389025646 +0000 UTC m=+41.701636804,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.436155 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa2adf4ad297\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf4ad297 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388703895 +0000 UTC m=+21.701315063,LastTimestamp:2026-03-14 09:11:14.389089368 +0000 UTC m=+41.701700526,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:11:38 crc kubenswrapper[4843]: E0314 09:11:38.444331 4843 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189caa2adf49a997\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 14 09:11:38 crc kubenswrapper[4843]: &Event{ObjectMeta:{kube-controller-manager-crc.189caa2adf49a997 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 14 09:11:38 crc kubenswrapper[4843]: body: Mar 14 09:11:38 crc kubenswrapper[4843]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:10:54.388627863 +0000 UTC m=+21.701239031,LastTimestamp:2026-03-14 09:11:24.388366189 +0000 UTC m=+51.700977327,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 14 09:11:38 crc kubenswrapper[4843]: > Mar 14 09:11:39 crc kubenswrapper[4843]: I0314 09:11:39.264630 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:40 crc kubenswrapper[4843]: I0314 09:11:40.264477 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:41 crc kubenswrapper[4843]: I0314 09:11:41.267455 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:41 crc kubenswrapper[4843]: I0314 09:11:41.387939 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:41 crc kubenswrapper[4843]: I0314 09:11:41.388540 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:41 crc kubenswrapper[4843]: I0314 09:11:41.394015 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:41 crc kubenswrapper[4843]: I0314 09:11:41.394073 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:41 crc kubenswrapper[4843]: I0314 09:11:41.394095 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:42 crc kubenswrapper[4843]: I0314 09:11:42.265972 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:43 crc kubenswrapper[4843]: I0314 09:11:43.265106 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:43 crc kubenswrapper[4843]: I0314 09:11:43.280418 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:43 crc kubenswrapper[4843]: I0314 09:11:43.281646 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:43 crc kubenswrapper[4843]: I0314 09:11:43.281675 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:43 crc kubenswrapper[4843]: I0314 09:11:43.281684 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:43 crc kubenswrapper[4843]: I0314 09:11:43.281702 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:43 crc kubenswrapper[4843]: E0314 09:11:43.287769 4843 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 14 09:11:43 crc kubenswrapper[4843]: E0314 09:11:43.291617 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 14 09:11:43 crc kubenswrapper[4843]: E0314 09:11:43.430777 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.161323 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.161479 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.162635 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.162689 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.162707 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.168927 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.266923 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.338841 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.340101 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.340195 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.340217 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.341412 4843 scope.go:117] "RemoveContainer" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.680296 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.682149 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.682614 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec"} Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.682702 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.683349 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.683373 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.683381 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.684142 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.684161 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:44 crc kubenswrapper[4843]: I0314 09:11:44.684169 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.266141 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.378656 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.398921 4843 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.698714 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.700034 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.702920 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" exitCode=255 Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.702985 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec"} Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.703038 4843 scope.go:117] "RemoveContainer" containerID="99ba3701c9339375883f021f55d37982759fdf28af9c191cd7ff6c5f5f912543" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.703257 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.704716 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.704761 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.704778 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:45 crc kubenswrapper[4843]: I0314 09:11:45.705590 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:11:45 crc kubenswrapper[4843]: E0314 09:11:45.705860 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:46 crc kubenswrapper[4843]: I0314 09:11:46.267484 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:46 crc kubenswrapper[4843]: I0314 09:11:46.706562 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 14 09:11:47 crc kubenswrapper[4843]: I0314 09:11:47.265145 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:48 crc kubenswrapper[4843]: I0314 09:11:48.264198 4843 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 14 09:11:48 crc kubenswrapper[4843]: I0314 09:11:48.339420 4843 csr.go:261] certificate signing request csr-zjds4 is approved, waiting to be issued Mar 14 09:11:48 crc kubenswrapper[4843]: I0314 09:11:48.346249 4843 csr.go:257] certificate signing request csr-zjds4 is issued Mar 14 09:11:48 crc kubenswrapper[4843]: I0314 09:11:48.383561 4843 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.121356 4843 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.248570 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.248782 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.250157 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.250195 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.250207 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.250902 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:11:49 crc kubenswrapper[4843]: E0314 09:11:49.251077 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.347363 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-05 12:59:17.090773156 +0000 UTC Mar 14 09:11:49 crc kubenswrapper[4843]: I0314 09:11:49.347426 4843 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7131h47m27.743352168s for next certificate rotation Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.288969 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.290603 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.290705 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.290733 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.291013 4843 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.298153 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.298434 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.299930 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.299975 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.299984 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.300732 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.300920 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.302239 4843 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.302720 4843 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.302753 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.307062 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.307119 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.307136 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.307160 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.307177 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:11:50Z","lastTransitionTime":"2026-03-14T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.327202 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.336016 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.336052 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.336067 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.336086 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.336102 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:11:50Z","lastTransitionTime":"2026-03-14T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.350886 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.361868 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.361911 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.361926 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.361946 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.361961 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:11:50Z","lastTransitionTime":"2026-03-14T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.375525 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.384244 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.384334 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.384351 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.384370 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:11:50 crc kubenswrapper[4843]: I0314 09:11:50.384386 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:11:50Z","lastTransitionTime":"2026-03-14T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.395765 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.395961 4843 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.396002 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.496550 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.596889 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.697019 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.798378 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:50 crc kubenswrapper[4843]: E0314 09:11:50.899526 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.001137 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.101893 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.202851 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.303351 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.403837 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.505158 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.605344 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.706122 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.807300 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:51 crc kubenswrapper[4843]: I0314 09:11:51.845225 4843 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 14 09:11:51 crc kubenswrapper[4843]: E0314 09:11:51.907647 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.008386 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.108930 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.209927 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.310373 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.410547 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.511455 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.612200 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.713212 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.813815 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:52 crc kubenswrapper[4843]: E0314 09:11:52.914433 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.014781 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.115108 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: I0314 09:11:53.193584 4843 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.216072 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.317008 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.417249 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.431601 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.518128 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.618906 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.720330 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.821375 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:53 crc kubenswrapper[4843]: E0314 09:11:53.922336 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.023055 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.123812 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.224024 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.324801 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.425972 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.526143 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.626841 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.727928 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.828348 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:54 crc kubenswrapper[4843]: E0314 09:11:54.929358 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.030401 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: I0314 09:11:55.035846 4843 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.130746 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.230906 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.331364 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.432342 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.533374 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.634167 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.735387 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.835605 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:55 crc kubenswrapper[4843]: E0314 09:11:55.936563 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.037348 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.137756 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.237881 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.338557 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.439402 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.540415 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.641439 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.741738 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.841927 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:56 crc kubenswrapper[4843]: E0314 09:11:56.942598 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.043257 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.143922 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.244512 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.345620 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.446565 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.547418 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.648333 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.748595 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.848958 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:57 crc kubenswrapper[4843]: E0314 09:11:57.949653 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.050077 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.150248 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.251031 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.351893 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.452989 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.553260 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.654110 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.755056 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.856145 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:58 crc kubenswrapper[4843]: E0314 09:11:58.956729 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.057489 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.158356 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.259345 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.359959 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.460695 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.561588 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.662409 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.763308 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.863842 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:11:59 crc kubenswrapper[4843]: E0314 09:11:59.964479 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.065545 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.166185 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.267354 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.367892 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.468011 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.477099 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.483342 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.483387 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.483399 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.483416 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.483429 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:00Z","lastTransitionTime":"2026-03-14T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.498830 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.503754 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.503813 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.503834 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.503857 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.503869 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:00Z","lastTransitionTime":"2026-03-14T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.519816 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.525873 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.526075 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.526204 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.526347 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.526454 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:00Z","lastTransitionTime":"2026-03-14T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.539810 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.550727 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.550765 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.550774 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.550792 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:00 crc kubenswrapper[4843]: I0314 09:12:00.550804 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:00Z","lastTransitionTime":"2026-03-14T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.561785 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"272dcccf-22b1-47a7-8807-43844aeee65b\\\",\\\"systemUUID\\\":\\\"7ad45a7c-2b2c-4e9f-b1c1-584bf6854f24\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.562035 4843 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.568751 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.669668 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.770725 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.871213 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:00 crc kubenswrapper[4843]: E0314 09:12:00.971809 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.072864 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.173891 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.274335 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.375123 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.476522 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.577385 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.677809 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.777966 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.878386 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:01 crc kubenswrapper[4843]: E0314 09:12:01.979370 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.080130 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.180294 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.283160 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: I0314 09:12:02.338653 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:12:02 crc kubenswrapper[4843]: I0314 09:12:02.340619 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:02 crc kubenswrapper[4843]: I0314 09:12:02.340672 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:02 crc kubenswrapper[4843]: I0314 09:12:02.340688 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:02 crc kubenswrapper[4843]: I0314 09:12:02.341629 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.341931 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.383747 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.484089 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.584820 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.684966 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.786046 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.886567 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:02 crc kubenswrapper[4843]: E0314 09:12:02.987201 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.087953 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.189234 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.290119 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.390945 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.432242 4843 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.491285 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.592092 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.692837 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.793505 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.893956 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:03 crc kubenswrapper[4843]: E0314 09:12:03.994876 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.095599 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.196581 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.297212 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: I0314 09:12:04.338308 4843 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 14 09:12:04 crc kubenswrapper[4843]: I0314 09:12:04.339520 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:04 crc kubenswrapper[4843]: I0314 09:12:04.339557 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:04 crc kubenswrapper[4843]: I0314 09:12:04.339570 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.398263 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.498851 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.599459 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.700132 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.800583 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:04 crc kubenswrapper[4843]: E0314 09:12:04.901386 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.001982 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.102252 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.203027 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.304168 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.404757 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.505045 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.606009 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.706762 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.807689 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:05 crc kubenswrapper[4843]: E0314 09:12:05.908526 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: E0314 09:12:06.009488 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: E0314 09:12:06.110193 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: E0314 09:12:06.211202 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: E0314 09:12:06.311821 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: E0314 09:12:06.412448 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: E0314 09:12:06.512817 4843 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.541224 4843 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.615647 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.615690 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.615708 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.615730 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.615747 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:06Z","lastTransitionTime":"2026-03-14T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.718212 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.718313 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.718338 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.718365 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.718386 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:06Z","lastTransitionTime":"2026-03-14T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.821373 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.821415 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.821432 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.821453 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.821470 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:06Z","lastTransitionTime":"2026-03-14T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.923557 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.923631 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.923652 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.923678 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:06 crc kubenswrapper[4843]: I0314 09:12:06.923696 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:06Z","lastTransitionTime":"2026-03-14T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.027323 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.027377 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.027400 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.027435 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.027459 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.131428 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.131503 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.131526 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.131561 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.131583 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.233674 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.233743 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.233765 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.233795 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.233824 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.300540 4843 apiserver.go:52] "Watching apiserver" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.305611 4843 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.306016 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.306633 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.306763 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.306878 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.307106 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.307143 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.307504 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.307525 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.307580 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.307639 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.309204 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.310112 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.310303 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.310329 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.310789 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.311163 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.311379 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.312085 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.313073 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.338255 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.338322 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.338337 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.338361 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.338377 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.353253 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.367767 4843 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.369401 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.385044 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.395844 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.409114 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418408 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418522 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418577 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418635 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418694 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418748 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418799 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418855 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418903 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418949 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.418995 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419044 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419096 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419149 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419200 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419253 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419342 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419404 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419456 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419507 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419558 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419614 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419672 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419727 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419778 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419834 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419885 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419932 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419979 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420032 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420077 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420125 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420174 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420220 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420358 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420409 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420456 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419039 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420523 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419073 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419486 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419485 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419562 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419802 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.419970 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420078 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420152 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420464 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420503 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420680 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420708 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420731 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420753 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420777 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420795 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420804 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420863 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420903 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420935 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.420968 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421003 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421037 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421074 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421107 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421137 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421167 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421198 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421228 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421260 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421323 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421352 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421377 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421404 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421426 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421448 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421470 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421491 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421513 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421536 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421568 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421593 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421615 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421639 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421662 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421683 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421706 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421733 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421759 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421782 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421803 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421823 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421843 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421895 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421922 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421953 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421976 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421999 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422021 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422046 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422074 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422105 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422133 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422158 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422181 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422202 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422225 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422246 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422290 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422312 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422592 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422618 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422639 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422661 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422683 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422706 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422729 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422750 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422776 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422800 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422824 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422846 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422869 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422891 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422912 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422935 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422957 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422981 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423003 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423026 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423048 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423070 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423102 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423130 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423153 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423183 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423212 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423234 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423258 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423392 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423418 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423505 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423530 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423556 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423580 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423603 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423865 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423891 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423915 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423937 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423959 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423985 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424008 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424031 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424051 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424076 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424110 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424136 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424156 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424184 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424207 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424258 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424300 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424324 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424347 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424375 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424406 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424428 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424452 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424486 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424514 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424535 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424569 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424611 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424638 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424674 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424708 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424738 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424772 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424810 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424843 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424877 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424910 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424944 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424977 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425009 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425045 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425096 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425134 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425172 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425208 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425250 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425346 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425383 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425418 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425453 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425488 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425524 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425556 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425588 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425622 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425654 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425687 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425725 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425763 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425795 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421135 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421214 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421356 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.421350 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422129 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422827 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.422579 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423834 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423861 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.423679 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424059 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424196 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424255 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424258 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424548 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424543 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424654 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424749 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.424988 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425192 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426262 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426323 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426419 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426358 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425407 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425459 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425714 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425833 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426511 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.425969 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426703 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426751 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426814 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426847 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426867 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426969 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427211 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427344 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427718 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427783 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427786 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426855 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427912 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427957 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428038 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428119 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428188 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428249 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428300 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428331 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428359 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428386 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428410 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428508 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.426362 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428598 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428618 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428633 4843 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428646 4843 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428660 4843 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428675 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428688 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428700 4843 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428712 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428726 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428740 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428754 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428768 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428780 4843 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428794 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428806 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428819 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428832 4843 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428848 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428867 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428886 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428900 4843 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428915 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428928 4843 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428941 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428955 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428969 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428984 4843 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.428997 4843 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429010 4843 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429027 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429040 4843 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429073 4843 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429104 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429123 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429139 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429156 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429175 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429195 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427749 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427949 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.427979 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429009 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429498 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.429887 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.430381 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.430489 4843 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.430550 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:07.930529088 +0000 UTC m=+95.243140226 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.430960 4843 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.431008 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:07.930997681 +0000 UTC m=+95.243608819 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.431510 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.431650 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.431882 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.430302 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.434363 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.434482 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.435232 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.435609 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.436387 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.436429 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.436723 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.436942 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.437157 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.437217 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.438861 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.442669 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.443139 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.444541 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.445117 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.445547 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.446063 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.446394 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.446878 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.447039 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.447321 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.447420 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.448145 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.448137 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.448257 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.448688 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.449073 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.449170 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.449595 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.449610 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.449767 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.450205 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.450380 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.450443 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.450845 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.450963 4843 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.451143 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.451187 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.451203 4843 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.451301 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:07.951255402 +0000 UTC m=+95.263866750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.451252 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.451465 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.451491 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.451859 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.451961 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.451994 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.452013 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.452029 4843 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.452091 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:07.952065687 +0000 UTC m=+95.264676835 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.452741 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.452805 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.454804 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.454883 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.454917 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.454933 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.454957 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.454974 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.455500 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.455580 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.455618 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.455920 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.456038 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.456360 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.456502 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.456857 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.456909 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.457350 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.457711 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.458087 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.458137 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.458814 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.459217 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.459672 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.459767 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.460108 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.460179 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.460622 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.461066 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.462504 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.462692 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.462917 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.463080 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:07.963045528 +0000 UTC m=+95.275656826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.463380 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.463505 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.463618 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.463991 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.464736 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465040 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465223 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465450 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465470 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465548 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465568 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465578 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.465861 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.466266 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.466972 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467066 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467235 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467365 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467435 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467491 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467584 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467632 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.467818 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.468115 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.468382 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.468587 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.468826 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.469300 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.469511 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.469683 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.469702 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.470099 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.470123 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.471192 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.471543 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.471756 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.471859 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.472774 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.473122 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.473437 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.473551 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.473949 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.473482 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.474346 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.476198 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.479230 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.479464 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.479477 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.479847 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.480379 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.480582 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.480655 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.480810 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.480943 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.480996 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481373 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481394 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481293 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481701 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481903 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481428 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481476 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.481638 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.482310 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.482697 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.482955 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.483426 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.483463 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.483620 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.483896 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.484150 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.500715 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.508903 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.510366 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.512938 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530026 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530082 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530128 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530141 4843 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530151 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530159 4843 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530168 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530180 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530190 4843 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530199 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530211 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530219 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530226 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530236 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530245 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530253 4843 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530261 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530269 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530293 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530302 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530311 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530319 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530327 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530335 4843 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530345 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530354 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530364 4843 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530373 4843 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530381 4843 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530390 4843 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530399 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530407 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530415 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530428 4843 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530439 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530447 4843 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530455 4843 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530464 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530472 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530481 4843 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530488 4843 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530496 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530505 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530513 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530521 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530529 4843 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530539 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530548 4843 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530559 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530566 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530574 4843 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530583 4843 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530594 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530604 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530615 4843 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530626 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530637 4843 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530647 4843 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530658 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530666 4843 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530675 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530683 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530654 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530747 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530693 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530885 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530920 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530948 4843 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530972 4843 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.530997 4843 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531019 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531040 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531065 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531091 4843 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531116 4843 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531144 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531170 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531193 4843 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531218 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531239 4843 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531255 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531309 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531333 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531357 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531381 4843 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531403 4843 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531594 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531615 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531637 4843 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531657 4843 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531680 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531703 4843 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531726 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531747 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531770 4843 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531798 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531821 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531844 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531867 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531891 4843 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531914 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531939 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531963 4843 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.531988 4843 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532012 4843 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532035 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532058 4843 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532081 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532104 4843 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532168 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532197 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532223 4843 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532250 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532308 4843 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532334 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532357 4843 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532382 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532406 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532430 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532455 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532482 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532506 4843 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532531 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532557 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532582 4843 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532606 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532629 4843 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532655 4843 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532680 4843 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532708 4843 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532729 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532752 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532775 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532798 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532822 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532846 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532870 4843 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532895 4843 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532939 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532968 4843 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.532990 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533012 4843 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533035 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533066 4843 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533089 4843 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533112 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533137 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533163 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533191 4843 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533215 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533238 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533262 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533330 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533353 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533376 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533400 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533422 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533445 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.533470 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.558309 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.558353 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.558366 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.558385 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.558400 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.625034 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.638732 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.645815 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 14 09:12:07 crc kubenswrapper[4843]: W0314 09:12:07.652787 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c0d264ab469de9f855d63bb8a8353cd53e7511937888452de556b808c13dc16f WatchSource:0}: Error finding container c0d264ab469de9f855d63bb8a8353cd53e7511937888452de556b808c13dc16f: Status 404 returned error can't find the container with id c0d264ab469de9f855d63bb8a8353cd53e7511937888452de556b808c13dc16f Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.665254 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.665309 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.665322 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.665352 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.665365 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.768004 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.768055 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.768066 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.768086 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.768097 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.797044 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c0d264ab469de9f855d63bb8a8353cd53e7511937888452de556b808c13dc16f"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.798588 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7d91334c14ac225cca2595c847d02e15d20f880f9641e38759118824aab975cf"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.799696 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"07f97e30ea3004d990a00478967ee8810dfb41b0afbb6c22969cdad93cfd28cf"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.870728 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.870767 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.870777 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.870793 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.870805 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.942538 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.942597 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.942697 4843 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.942752 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:08.942738799 +0000 UTC m=+96.255349927 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.942789 4843 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: E0314 09:12:07.942808 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:08.942803141 +0000 UTC m=+96.255414269 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.973805 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.973845 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.973855 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.973873 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:07 crc kubenswrapper[4843]: I0314 09:12:07.973886 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:07Z","lastTransitionTime":"2026-03-14T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.043222 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.043360 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.043405 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043440 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:09.043402944 +0000 UTC m=+96.356014082 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043559 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043584 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043598 4843 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043601 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043636 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043648 4843 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043656 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:09.043637651 +0000 UTC m=+96.356248969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.043727 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:09.043708753 +0000 UTC m=+96.356319881 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.076994 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.077051 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.077065 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.077082 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.077101 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.178817 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.178859 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.178871 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.178889 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.178902 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.281107 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.281144 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.281152 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.281166 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.281176 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.382898 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.382927 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.382935 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.382948 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.382958 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.486109 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.486183 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.486192 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.486254 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.486287 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.589257 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.589356 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.589372 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.589397 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.589414 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.693189 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.693249 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.693298 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.693325 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.693342 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.796805 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.796847 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.796857 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.796873 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.796884 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.804012 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2d154eac00a21dd3ba5dea1ff0e5f7c23e57b4ac8cad803c625888b3a7dafa50"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.804084 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"10ac9aeb202c8761b8359f9b53c0666bb2835e711834e98956b2d4fef34bc257"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.805422 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5a6acdfadbac9a6f2b8d88edf8593671e9383a77b2d2724898478b19f3ce293b"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.827815 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d154eac00a21dd3ba5dea1ff0e5f7c23e57b4ac8cad803c625888b3a7dafa50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-14T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ac9aeb202c8761b8359f9b53c0666bb2835e711834e98956b2d4fef34bc257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-14T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.848497 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.862307 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.879921 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.899685 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.900028 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.900172 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.900355 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.900487 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:08Z","lastTransitionTime":"2026-03-14T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.904333 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.920257 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.937666 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d154eac00a21dd3ba5dea1ff0e5f7c23e57b4ac8cad803c625888b3a7dafa50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-14T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10ac9aeb202c8761b8359f9b53c0666bb2835e711834e98956b2d4fef34bc257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-14T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.951445 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.951518 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.951994 4843 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.952112 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:10.952078469 +0000 UTC m=+98.264689597 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.952518 4843 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: E0314 09:12:08.952603 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:10.952588275 +0000 UTC m=+98.265199403 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.956369 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a6acdfadbac9a6f2b8d88edf8593671e9383a77b2d2724898478b19f3ce293b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-14T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.973735 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:08 crc kubenswrapper[4843]: I0314 09:12:08.988623 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.001849 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.002957 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.003104 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.003230 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.003482 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.003655 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.014924 4843 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-14T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-14T09:12:09Z is after 2025-08-24T17:21:41Z" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.052444 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.052654 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:11.052618991 +0000 UTC m=+98.365230119 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.053139 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.053289 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.053339 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.053391 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.053681 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.053716 4843 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.053792 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:11.053769956 +0000 UTC m=+98.366381094 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.053581 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.054001 4843 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.054155 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:11.054130676 +0000 UTC m=+98.366741814 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.106856 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.107320 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.107520 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.107678 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.107817 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.210957 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.210995 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.211006 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.211020 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.211030 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.313993 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.314306 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.314384 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.314492 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.314612 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.338707 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.338715 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.339465 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.338791 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.339165 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:09 crc kubenswrapper[4843]: E0314 09:12:09.339757 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.343681 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.344478 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.346233 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.347215 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.348666 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.349453 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.350485 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.352014 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.353022 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.354492 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.355259 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.356880 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.357752 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.358593 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.359986 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.360538 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.361318 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.361906 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.362782 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.363610 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.364257 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.366610 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.367206 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.368746 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.369422 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.371014 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.371960 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.373259 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.374223 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.375484 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.376159 4843 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.376340 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.378697 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.379889 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.380495 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.382930 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.384440 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.385222 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.386758 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.387713 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.388944 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.389914 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.391322 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.391910 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.392806 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.393457 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.394369 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.395124 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.395951 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.396450 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.397254 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.397800 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.398359 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.399179 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.417007 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.417031 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.417039 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.417052 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.417061 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.519211 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.519241 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.519250 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.519264 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.519285 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.621846 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.621897 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.621908 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.621926 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.621938 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.724543 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.724579 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.724590 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.724604 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.724614 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.827146 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.827205 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.827216 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.827238 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.827253 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.930571 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.930618 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.930626 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.930640 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:09 crc kubenswrapper[4843]: I0314 09:12:09.930649 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:09Z","lastTransitionTime":"2026-03-14T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.033800 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.033842 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.033858 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.033879 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.033896 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.137039 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.137087 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.137099 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.137120 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.137131 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.240175 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.240266 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.240323 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.240354 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.240378 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.342750 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.342786 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.342797 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.342814 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.342826 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.446096 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.446141 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.446156 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.446177 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.446195 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.549068 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.549100 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.549109 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.549123 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.549137 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.651590 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.651634 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.651644 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.651666 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.651676 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.680786 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.680837 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.680849 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.680867 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.680882 4843 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-14T09:12:10Z","lastTransitionTime":"2026-03-14T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.813813 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4894022aebed86aa91fb06c26d2ae3dddcb12509022787f1606d46fcbb434dfc"} Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.970796 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:10 crc kubenswrapper[4843]: I0314 09:12:10.970864 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:10 crc kubenswrapper[4843]: E0314 09:12:10.970928 4843 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:10 crc kubenswrapper[4843]: E0314 09:12:10.970968 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:14.970955726 +0000 UTC m=+102.283566854 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:10 crc kubenswrapper[4843]: E0314 09:12:10.971091 4843 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:10 crc kubenswrapper[4843]: E0314 09:12:10.971189 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:14.971166332 +0000 UTC m=+102.283777490 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.071705 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.071785 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.071841 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.071959 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.071979 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.071986 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.071989 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:15.07194348 +0000 UTC m=+102.384554638 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.072019 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.072051 4843 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.072119 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:15.072096435 +0000 UTC m=+102.384707603 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.071992 4843 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.072223 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:15.072208828 +0000 UTC m=+102.384819996 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.328480 4843 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.338588 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.338669 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.338878 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.338895 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.339100 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:11 crc kubenswrapper[4843]: E0314 09:12:11.339202 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:11 crc kubenswrapper[4843]: I0314 09:12:11.340779 4843 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 14 09:12:13 crc kubenswrapper[4843]: I0314 09:12:13.338059 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:13 crc kubenswrapper[4843]: I0314 09:12:13.338075 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:13 crc kubenswrapper[4843]: I0314 09:12:13.338161 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:13 crc kubenswrapper[4843]: E0314 09:12:13.338417 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:13 crc kubenswrapper[4843]: E0314 09:12:13.338565 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:13 crc kubenswrapper[4843]: E0314 09:12:13.338798 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.011140 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.011219 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.011386 4843 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.011453 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:23.01143472 +0000 UTC m=+110.324045858 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.011882 4843 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.011929 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:23.011915544 +0000 UTC m=+110.324526692 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.020000 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2sxrf"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.020389 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.022965 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.023615 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.026326 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.042692 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-sd9qb"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.043059 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gwd22"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.043357 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-p8t7w"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.044135 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.044628 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.045423 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.049709 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.051908 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.052234 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.052376 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.052770 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.053331 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.053548 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.053693 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.053853 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.065212 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.066843 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.067136 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.068332 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vfpmm"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.069861 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.072245 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.072746 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.073195 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.073660 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.073728 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.073964 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.074968 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.112605 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.112707 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.112757 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7kvq\" (UniqueName: \"kubernetes.io/projected/3762367a-eb3b-4992-9f00-5dd12b8ff0b1-kube-api-access-k7kvq\") pod \"node-resolver-2sxrf\" (UID: \"3762367a-eb3b-4992-9f00-5dd12b8ff0b1\") " pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.112790 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.112811 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3762367a-eb3b-4992-9f00-5dd12b8ff0b1-hosts-file\") pod \"node-resolver-2sxrf\" (UID: \"3762367a-eb3b-4992-9f00-5dd12b8ff0b1\") " pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.112894 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:23.112878358 +0000 UTC m=+110.425489486 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113007 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113025 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113039 4843 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113071 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:23.113064204 +0000 UTC m=+110.425675332 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113417 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113446 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113459 4843 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.113493 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:23.113482837 +0000 UTC m=+110.426093965 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.142400 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.143061 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.144854 4843 reflector.go:561] object-"openshift-cluster-version"/"default-dockercfg-gxtc4": failed to list *v1.Secret: secrets "default-dockercfg-gxtc4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.144943 4843 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"default-dockercfg-gxtc4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-gxtc4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.145248 4843 reflector.go:561] object-"openshift-cluster-version"/"cluster-version-operator-serving-cert": failed to list *v1.Secret: secrets "cluster-version-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.145296 4843 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cluster-version-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.145351 4843 reflector.go:561] object-"openshift-cluster-version"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.145371 4843 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.145419 4843 reflector.go:561] object-"openshift-cluster-version"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.145470 4843 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213204 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-script-lib\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213241 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f55462c-aaf9-41fa-bae6-2f7fa118841b-proxy-tls\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213258 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213293 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-kubelet\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213364 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3762367a-eb3b-4992-9f00-5dd12b8ff0b1-hosts-file\") pod \"node-resolver-2sxrf\" (UID: \"3762367a-eb3b-4992-9f00-5dd12b8ff0b1\") " pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213448 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-cnibin\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213509 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-socket-dir-parent\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213531 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-bin\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213528 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3762367a-eb3b-4992-9f00-5dd12b8ff0b1-hosts-file\") pod \"node-resolver-2sxrf\" (UID: \"3762367a-eb3b-4992-9f00-5dd12b8ff0b1\") " pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213551 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-netd\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213600 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8f55462c-aaf9-41fa-bae6-2f7fa118841b-rootfs\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213665 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-kubelet\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213741 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-conf-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213786 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-ovn-kubernetes\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213816 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovn-node-metrics-cert\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213853 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-netns\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213871 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-var-lib-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213892 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-daemon-config\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213939 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-os-release\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213975 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-node-log\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.213994 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqwqh\" (UniqueName: \"kubernetes.io/projected/986f5a7d-244f-4068-bd4d-949b7afc1e16-kube-api-access-gqwqh\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214029 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-cnibin\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214060 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-system-cni-dir\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214105 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32cf92ff-937a-41d1-b27a-127bd550a1eb-cni-binary-copy\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214127 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-cni-multus\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214149 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-systemd-units\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214184 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-log-socket\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214201 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-k8s-cni-cncf-io\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214251 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-slash\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214289 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-ovn\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214308 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-system-cni-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214374 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/da3d426b-76cf-40bd-8368-f52b1f74e841-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214400 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-cni-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214419 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-cni-bin\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214441 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7kvq\" (UniqueName: \"kubernetes.io/projected/3762367a-eb3b-4992-9f00-5dd12b8ff0b1-kube-api-access-k7kvq\") pod \"node-resolver-2sxrf\" (UID: \"3762367a-eb3b-4992-9f00-5dd12b8ff0b1\") " pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214474 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214496 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f55462c-aaf9-41fa-bae6-2f7fa118841b-mcd-auth-proxy-config\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214515 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7blw8\" (UniqueName: \"kubernetes.io/projected/32cf92ff-937a-41d1-b27a-127bd550a1eb-kube-api-access-7blw8\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214537 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq94m\" (UniqueName: \"kubernetes.io/projected/da3d426b-76cf-40bd-8368-f52b1f74e841-kube-api-access-jq94m\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214562 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-etc-kubernetes\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214583 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214600 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-config\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214617 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-os-release\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214637 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/da3d426b-76cf-40bd-8368-f52b1f74e841-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214656 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-env-overrides\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214671 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-netns\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214688 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-multus-certs\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214741 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-systemd\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214769 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-etc-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214792 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppvzs\" (UniqueName: \"kubernetes.io/projected/8f55462c-aaf9-41fa-bae6-2f7fa118841b-kube-api-access-ppvzs\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.214820 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-hostroot\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.232060 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7kvq\" (UniqueName: \"kubernetes.io/projected/3762367a-eb3b-4992-9f00-5dd12b8ff0b1-kube-api-access-k7kvq\") pod \"node-resolver-2sxrf\" (UID: \"3762367a-eb3b-4992-9f00-5dd12b8ff0b1\") " pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.271944 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xdc5m"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.272383 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.274477 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.275576 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.275696 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.275779 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316075 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-node-log\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316191 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqwqh\" (UniqueName: \"kubernetes.io/projected/986f5a7d-244f-4068-bd4d-949b7afc1e16-kube-api-access-gqwqh\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316227 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-node-log\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316616 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-cnibin\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316647 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-system-cni-dir\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316743 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32cf92ff-937a-41d1-b27a-127bd550a1eb-cni-binary-copy\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316815 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-cni-multus\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316711 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-cnibin\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316746 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-system-cni-dir\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316914 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-systemd-units\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317122 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-log-socket\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317201 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-log-socket\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317207 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-k8s-cni-cncf-io\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317264 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d155422d-29df-4707-8b98-524a7dee63e8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316970 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-systemd-units\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317329 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-slash\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317360 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-ovn\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317363 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-slash\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317385 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-system-cni-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317404 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-ovn\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317455 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/da3d426b-76cf-40bd-8368-f52b1f74e841-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317487 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-cni-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317511 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-cni-bin\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317541 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317567 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f55462c-aaf9-41fa-bae6-2f7fa118841b-mcd-auth-proxy-config\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317590 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7blw8\" (UniqueName: \"kubernetes.io/projected/32cf92ff-937a-41d1-b27a-127bd550a1eb-kube-api-access-7blw8\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317646 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq94m\" (UniqueName: \"kubernetes.io/projected/da3d426b-76cf-40bd-8368-f52b1f74e841-kube-api-access-jq94m\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317669 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-etc-kubernetes\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317693 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317718 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-config\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317740 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-os-release\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317762 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/da3d426b-76cf-40bd-8368-f52b1f74e841-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317791 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32cf92ff-937a-41d1-b27a-127bd550a1eb-cni-binary-copy\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317785 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-env-overrides\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317867 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-netns\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317887 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-multus-certs\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317907 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-systemd\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317939 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-etc-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317961 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d155422d-29df-4707-8b98-524a7dee63e8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317982 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d155422d-29df-4707-8b98-524a7dee63e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318015 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d155422d-29df-4707-8b98-524a7dee63e8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318033 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppvzs\" (UniqueName: \"kubernetes.io/projected/8f55462c-aaf9-41fa-bae6-2f7fa118841b-kube-api-access-ppvzs\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318050 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-hostroot\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318071 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-script-lib\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318105 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f55462c-aaf9-41fa-bae6-2f7fa118841b-proxy-tls\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318132 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318153 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-kubelet\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318199 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-cnibin\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318220 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-socket-dir-parent\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318239 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-bin\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318293 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-netd\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318301 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/da3d426b-76cf-40bd-8368-f52b1f74e841-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318317 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8f55462c-aaf9-41fa-bae6-2f7fa118841b-rootfs\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.316924 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-cni-multus\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318368 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-kubelet\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318407 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-kubelet\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318447 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-conf-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318470 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-netns\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.317457 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-system-cni-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318500 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-multus-certs\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318486 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-ovn-kubernetes\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318531 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-ovn-kubernetes\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318538 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovn-node-metrics-cert\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318559 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-cni-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318588 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-netns\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318590 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-conf-dir\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318601 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-env-overrides\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318610 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318632 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-cni-bin\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318565 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-netns\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318850 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-etc-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318880 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-hostroot\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318896 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-netd\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318913 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8f55462c-aaf9-41fa-bae6-2f7fa118841b-rootfs\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318938 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.318955 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-var-lib-kubelet\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319066 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-host-run-k8s-cni-cncf-io\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319166 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-systemd\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319211 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-cnibin\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319236 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-bin\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319303 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-var-lib-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319332 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-daemon-config\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319361 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-os-release\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319390 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d155422d-29df-4707-8b98-524a7dee63e8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319456 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-var-lib-openvswitch\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319535 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/da3d426b-76cf-40bd-8368-f52b1f74e841-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319599 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-os-release\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319647 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-socket-dir-parent\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319702 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-os-release\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319726 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/32cf92ff-937a-41d1-b27a-127bd550a1eb-etc-kubernetes\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.319832 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-script-lib\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.320232 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/32cf92ff-937a-41d1-b27a-127bd550a1eb-multus-daemon-config\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.320447 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f55462c-aaf9-41fa-bae6-2f7fa118841b-mcd-auth-proxy-config\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.320483 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-config\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.320840 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/da3d426b-76cf-40bd-8368-f52b1f74e841-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.323504 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f55462c-aaf9-41fa-bae6-2f7fa118841b-proxy-tls\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.331155 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovn-node-metrics-cert\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.335950 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2sxrf" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.339095 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.339322 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.339401 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.339579 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.339679 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.339731 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.347679 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqwqh\" (UniqueName: \"kubernetes.io/projected/986f5a7d-244f-4068-bd4d-949b7afc1e16-kube-api-access-gqwqh\") pod \"ovnkube-node-vfpmm\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.350054 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq94m\" (UniqueName: \"kubernetes.io/projected/da3d426b-76cf-40bd-8368-f52b1f74e841-kube-api-access-jq94m\") pod \"multus-additional-cni-plugins-p8t7w\" (UID: \"da3d426b-76cf-40bd-8368-f52b1f74e841\") " pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.350503 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7blw8\" (UniqueName: \"kubernetes.io/projected/32cf92ff-937a-41d1-b27a-127bd550a1eb-kube-api-access-7blw8\") pod \"multus-sd9qb\" (UID: \"32cf92ff-937a-41d1-b27a-127bd550a1eb\") " pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.352317 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppvzs\" (UniqueName: \"kubernetes.io/projected/8f55462c-aaf9-41fa-bae6-2f7fa118841b-kube-api-access-ppvzs\") pod \"machine-config-daemon-gwd22\" (UID: \"8f55462c-aaf9-41fa-bae6-2f7fa118841b\") " pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.354296 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3762367a_eb3b_4992_9f00_5dd12b8ff0b1.slice/crio-bf65ea3ceb67c87f552cb617844e899020e68227686a094ae8077ce4d02e032d WatchSource:0}: Error finding container bf65ea3ceb67c87f552cb617844e899020e68227686a094ae8077ce4d02e032d: Status 404 returned error can't find the container with id bf65ea3ceb67c87f552cb617844e899020e68227686a094ae8077ce4d02e032d Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.363583 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.365540 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.376842 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sd9qb" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.381266 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda3d426b_76cf_40bd_8368_f52b1f74e841.slice/crio-afebafbe4db49545666d88d31788aace9eab08c8001cb175f252c22c1309f9bd WatchSource:0}: Error finding container afebafbe4db49545666d88d31788aace9eab08c8001cb175f252c22c1309f9bd: Status 404 returned error can't find the container with id afebafbe4db49545666d88d31788aace9eab08c8001cb175f252c22c1309f9bd Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.386139 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.390245 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32cf92ff_937a_41d1_b27a_127bd550a1eb.slice/crio-06cc7b49abed7b4f78407e004d84dc152d03afe7c728141cebb5ea7d40493c06 WatchSource:0}: Error finding container 06cc7b49abed7b4f78407e004d84dc152d03afe7c728141cebb5ea7d40493c06: Status 404 returned error can't find the container with id 06cc7b49abed7b4f78407e004d84dc152d03afe7c728141cebb5ea7d40493c06 Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.392901 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420723 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d155422d-29df-4707-8b98-524a7dee63e8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420779 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d34f055d-2583-40a9-948d-3abc010b3996-serviceca\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420820 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d155422d-29df-4707-8b98-524a7dee63e8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420868 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d155422d-29df-4707-8b98-524a7dee63e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420891 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d155422d-29df-4707-8b98-524a7dee63e8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420915 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d34f055d-2583-40a9-948d-3abc010b3996-host\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.420884 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d155422d-29df-4707-8b98-524a7dee63e8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.421544 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hmmj\" (UniqueName: \"kubernetes.io/projected/d34f055d-2583-40a9-948d-3abc010b3996-kube-api-access-8hmmj\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.421590 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d155422d-29df-4707-8b98-524a7dee63e8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.421670 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d155422d-29df-4707-8b98-524a7dee63e8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.430457 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod986f5a7d_244f_4068_bd4d_949b7afc1e16.slice/crio-63d446a30e03b814adb1e4107ae699b45215b3e3da41ad7fb3a55c457707f1a6 WatchSource:0}: Error finding container 63d446a30e03b814adb1e4107ae699b45215b3e3da41ad7fb3a55c457707f1a6: Status 404 returned error can't find the container with id 63d446a30e03b814adb1e4107ae699b45215b3e3da41ad7fb3a55c457707f1a6 Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.472759 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.473428 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.475072 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.475721 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.493213 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2zndz"] Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.495447 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.495848 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.506116 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=0.506097903 podStartE2EDuration="506.097903ms" podCreationTimestamp="2026-03-14 09:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:15.505048471 +0000 UTC m=+102.817659599" watchObservedRunningTime="2026-03-14 09:12:15.506097903 +0000 UTC m=+102.818709031" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.522965 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d34f055d-2583-40a9-948d-3abc010b3996-host\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.523007 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hmmj\" (UniqueName: \"kubernetes.io/projected/d34f055d-2583-40a9-948d-3abc010b3996-kube-api-access-8hmmj\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.523043 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d34f055d-2583-40a9-948d-3abc010b3996-serviceca\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.523117 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d34f055d-2583-40a9-948d-3abc010b3996-host\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.524191 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d34f055d-2583-40a9-948d-3abc010b3996-serviceca\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.538736 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hmmj\" (UniqueName: \"kubernetes.io/projected/d34f055d-2583-40a9-948d-3abc010b3996-kube-api-access-8hmmj\") pod \"node-ca-xdc5m\" (UID: \"d34f055d-2583-40a9-948d-3abc010b3996\") " pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.584090 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xdc5m" Mar 14 09:12:15 crc kubenswrapper[4843]: W0314 09:12:15.596819 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd34f055d_2583_40a9_948d_3abc010b3996.slice/crio-29f30266a9179c2c578edf8e73c94beb9a490927af01f84265d1432e50e8dce2 WatchSource:0}: Error finding container 29f30266a9179c2c578edf8e73c94beb9a490927af01f84265d1432e50e8dce2: Status 404 returned error can't find the container with id 29f30266a9179c2c578edf8e73c94beb9a490927af01f84265d1432e50e8dce2 Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.623623 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/70117c20-1b22-4e7e-a663-1665e5b64a0f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.623672 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/70117c20-1b22-4e7e-a663-1665e5b64a0f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.623740 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkb29\" (UniqueName: \"kubernetes.io/projected/70117c20-1b22-4e7e-a663-1665e5b64a0f-kube-api-access-nkb29\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.623764 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/70117c20-1b22-4e7e-a663-1665e5b64a0f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.623790 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rspbk\" (UniqueName: \"kubernetes.io/projected/ca968925-6d0f-46b7-b12b-7c617faabd3c-kube-api-access-rspbk\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.623837 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.725502 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/70117c20-1b22-4e7e-a663-1665e5b64a0f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.725550 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/70117c20-1b22-4e7e-a663-1665e5b64a0f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.725599 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkb29\" (UniqueName: \"kubernetes.io/projected/70117c20-1b22-4e7e-a663-1665e5b64a0f-kube-api-access-nkb29\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.725616 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/70117c20-1b22-4e7e-a663-1665e5b64a0f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.725635 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rspbk\" (UniqueName: \"kubernetes.io/projected/ca968925-6d0f-46b7-b12b-7c617faabd3c-kube-api-access-rspbk\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.725661 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.725757 4843 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: E0314 09:12:15.725804 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs podName:ca968925-6d0f-46b7-b12b-7c617faabd3c nodeName:}" failed. No retries permitted until 2026-03-14 09:12:16.225790507 +0000 UTC m=+103.538401635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs") pod "network-metrics-daemon-2zndz" (UID: "ca968925-6d0f-46b7-b12b-7c617faabd3c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.730957 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/70117c20-1b22-4e7e-a663-1665e5b64a0f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.731107 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/70117c20-1b22-4e7e-a663-1665e5b64a0f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.731241 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/70117c20-1b22-4e7e-a663-1665e5b64a0f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.745131 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rspbk\" (UniqueName: \"kubernetes.io/projected/ca968925-6d0f-46b7-b12b-7c617faabd3c-kube-api-access-rspbk\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.747250 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkb29\" (UniqueName: \"kubernetes.io/projected/70117c20-1b22-4e7e-a663-1665e5b64a0f-kube-api-access-nkb29\") pod \"ovnkube-control-plane-749d76644c-lqxth\" (UID: \"70117c20-1b22-4e7e-a663-1665e5b64a0f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.837517 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.843414 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" exitCode=0 Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.843522 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.843570 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"63d446a30e03b814adb1e4107ae699b45215b3e3da41ad7fb3a55c457707f1a6"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.847657 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"dff53e211e9e21890b5d4be51c6077ef6982197a831e587b29b06b6212b691f2"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.847684 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"1eec10ad2604b241d08a4efa3556d36326f8f05b555f3f2dc5588bc1ea55449f"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.847694 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"f4c9ee99d07ba7d0c93ff49e22adb9cff9bd8b348e7577645c1ba31f0ce75835"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.850911 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xdc5m" event={"ID":"d34f055d-2583-40a9-948d-3abc010b3996","Type":"ContainerStarted","Data":"fb07e16d1f296e34970f53e4c4d2efb50a5a52566bc779c906b714253be15ea4"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.850934 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xdc5m" event={"ID":"d34f055d-2583-40a9-948d-3abc010b3996","Type":"ContainerStarted","Data":"29f30266a9179c2c578edf8e73c94beb9a490927af01f84265d1432e50e8dce2"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.852401 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sd9qb" event={"ID":"32cf92ff-937a-41d1-b27a-127bd550a1eb","Type":"ContainerStarted","Data":"e93284b6b38fee95cbefc6bd220dc1328863323ce552a6eb6bed85a982184a8c"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.852444 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sd9qb" event={"ID":"32cf92ff-937a-41d1-b27a-127bd550a1eb","Type":"ContainerStarted","Data":"06cc7b49abed7b4f78407e004d84dc152d03afe7c728141cebb5ea7d40493c06"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.854856 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerStarted","Data":"b0f333da12fadaa0e6875aef64f77cba94b0dbc17f186f0f8d6166b307fa6cdb"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.854922 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerStarted","Data":"afebafbe4db49545666d88d31788aace9eab08c8001cb175f252c22c1309f9bd"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.858446 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2sxrf" event={"ID":"3762367a-eb3b-4992-9f00-5dd12b8ff0b1","Type":"ContainerStarted","Data":"4b81cd85952e7dc896ea45c27dbaea99569186e98f9b299ea9af22f7bbcb1a0f"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.858480 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2sxrf" event={"ID":"3762367a-eb3b-4992-9f00-5dd12b8ff0b1","Type":"ContainerStarted","Data":"bf65ea3ceb67c87f552cb617844e899020e68227686a094ae8077ce4d02e032d"} Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.902041 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-sd9qb" podStartSLOduration=29.9020141 podStartE2EDuration="29.9020141s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:15.900558386 +0000 UTC m=+103.213169514" watchObservedRunningTime="2026-03-14 09:12:15.9020141 +0000 UTC m=+103.214625268" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.948979 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podStartSLOduration=29.948956025 podStartE2EDuration="29.948956025s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:15.922231849 +0000 UTC m=+103.234842997" watchObservedRunningTime="2026-03-14 09:12:15.948956025 +0000 UTC m=+103.261567153" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.966658 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2sxrf" podStartSLOduration=29.966637888 podStartE2EDuration="29.966637888s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:15.966432622 +0000 UTC m=+103.279043760" watchObservedRunningTime="2026-03-14 09:12:15.966637888 +0000 UTC m=+103.279249016" Mar 14 09:12:15 crc kubenswrapper[4843]: I0314 09:12:15.984557 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xdc5m" podStartSLOduration=29.984538078 podStartE2EDuration="29.984538078s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:15.984523928 +0000 UTC m=+103.297135066" watchObservedRunningTime="2026-03-14 09:12:15.984538078 +0000 UTC m=+103.297149206" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.037446 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.042577 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d155422d-29df-4707-8b98-524a7dee63e8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.228061 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.230716 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:16 crc kubenswrapper[4843]: E0314 09:12:16.231162 4843 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:16 crc kubenswrapper[4843]: E0314 09:12:16.231292 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs podName:ca968925-6d0f-46b7-b12b-7c617faabd3c nodeName:}" failed. No retries permitted until 2026-03-14 09:12:17.231242846 +0000 UTC m=+104.543853984 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs") pod "network-metrics-daemon-2zndz" (UID: "ca968925-6d0f-46b7-b12b-7c617faabd3c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.238988 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d155422d-29df-4707-8b98-524a7dee63e8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:16 crc kubenswrapper[4843]: E0314 09:12:16.421744 4843 secret.go:188] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 14 09:12:16 crc kubenswrapper[4843]: E0314 09:12:16.421862 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d155422d-29df-4707-8b98-524a7dee63e8-serving-cert podName:d155422d-29df-4707-8b98-524a7dee63e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:16.921834701 +0000 UTC m=+104.234445849 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d155422d-29df-4707-8b98-524a7dee63e8-serving-cert") pod "cluster-version-operator-5c965bbfc6-h6gk2" (UID: "d155422d-29df-4707-8b98-524a7dee63e8") : failed to sync secret cache: timed out waiting for the condition Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.566828 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.685422 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.864017 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.864574 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.864596 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.866543 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" event={"ID":"70117c20-1b22-4e7e-a663-1665e5b64a0f","Type":"ContainerStarted","Data":"5a5f906f5531d98dede671dbf95f6e01d6f0d39f440a678d0af8803b8bfc4cd3"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.866571 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" event={"ID":"70117c20-1b22-4e7e-a663-1665e5b64a0f","Type":"ContainerStarted","Data":"b55d00bc7a84e26118afdd6c51fa0790524327c182cb19d18bb910aa5f1bb448"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.866587 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" event={"ID":"70117c20-1b22-4e7e-a663-1665e5b64a0f","Type":"ContainerStarted","Data":"ca5e455c855ba1a55df1bcc5d31dd40c24232cb7dfeee4deb48bb3a0ab17a7ab"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.868056 4843 generic.go:334] "Generic (PLEG): container finished" podID="da3d426b-76cf-40bd-8368-f52b1f74e841" containerID="b0f333da12fadaa0e6875aef64f77cba94b0dbc17f186f0f8d6166b307fa6cdb" exitCode=0 Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.868111 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerDied","Data":"b0f333da12fadaa0e6875aef64f77cba94b0dbc17f186f0f8d6166b307fa6cdb"} Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.938985 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d155422d-29df-4707-8b98-524a7dee63e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.947502 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d155422d-29df-4707-8b98-524a7dee63e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h6gk2\" (UID: \"d155422d-29df-4707-8b98-524a7dee63e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:16 crc kubenswrapper[4843]: I0314 09:12:16.955718 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.242433 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.242645 4843 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.242739 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs podName:ca968925-6d0f-46b7-b12b-7c617faabd3c nodeName:}" failed. No retries permitted until 2026-03-14 09:12:19.242718011 +0000 UTC m=+106.555329129 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs") pod "network-metrics-daemon-2zndz" (UID: "ca968925-6d0f-46b7-b12b-7c617faabd3c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.342449 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.342562 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.342596 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.342467 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.342449 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.343040 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.343110 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.343155 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.359109 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.359423 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.360550 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.876057 4843 generic.go:334] "Generic (PLEG): container finished" podID="da3d426b-76cf-40bd-8368-f52b1f74e841" containerID="4a6ef245ca02649546c0b31bee7ef7ccb1726f1a58322232d1f97a9409f8c403" exitCode=0 Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.876107 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerDied","Data":"4a6ef245ca02649546c0b31bee7ef7ccb1726f1a58322232d1f97a9409f8c403"} Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.879034 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" event={"ID":"d155422d-29df-4707-8b98-524a7dee63e8","Type":"ContainerStarted","Data":"e4741ca400f9cc35b61eafb7677559f074e2686533bd7afa55616164cb85b66b"} Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.879057 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" event={"ID":"d155422d-29df-4707-8b98-524a7dee63e8","Type":"ContainerStarted","Data":"0aa6069fc89b6319547f70ca69d6720442b10c196bceb5421bd232a5ee525edf"} Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.886631 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:12:17 crc kubenswrapper[4843]: E0314 09:12:17.886877 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.887108 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.887160 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.887177 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.953618 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h6gk2" podStartSLOduration=31.953586332 podStartE2EDuration="31.953586332s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:17.951968563 +0000 UTC m=+105.264579701" watchObservedRunningTime="2026-03-14 09:12:17.953586332 +0000 UTC m=+105.266197450" Mar 14 09:12:17 crc kubenswrapper[4843]: I0314 09:12:17.972114 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqxth" podStartSLOduration=30.972087179 podStartE2EDuration="30.972087179s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:17.97177888 +0000 UTC m=+105.284390008" watchObservedRunningTime="2026-03-14 09:12:17.972087179 +0000 UTC m=+105.284698317" Mar 14 09:12:18 crc kubenswrapper[4843]: I0314 09:12:18.893925 4843 generic.go:334] "Generic (PLEG): container finished" podID="da3d426b-76cf-40bd-8368-f52b1f74e841" containerID="ab01e7de2361566619017b9ce82c16333e23da62804f57d45dc1b8f833b0ff4b" exitCode=0 Mar 14 09:12:18 crc kubenswrapper[4843]: I0314 09:12:18.894007 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerDied","Data":"ab01e7de2361566619017b9ce82c16333e23da62804f57d45dc1b8f833b0ff4b"} Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.264984 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:19 crc kubenswrapper[4843]: E0314 09:12:19.265158 4843 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:19 crc kubenswrapper[4843]: E0314 09:12:19.265223 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs podName:ca968925-6d0f-46b7-b12b-7c617faabd3c nodeName:}" failed. No retries permitted until 2026-03-14 09:12:23.265203805 +0000 UTC m=+110.577814933 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs") pod "network-metrics-daemon-2zndz" (UID: "ca968925-6d0f-46b7-b12b-7c617faabd3c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.338599 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.338653 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.338611 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.338606 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:19 crc kubenswrapper[4843]: E0314 09:12:19.338840 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:19 crc kubenswrapper[4843]: E0314 09:12:19.338959 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:19 crc kubenswrapper[4843]: E0314 09:12:19.339032 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:19 crc kubenswrapper[4843]: E0314 09:12:19.339080 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.899621 4843 generic.go:334] "Generic (PLEG): container finished" podID="da3d426b-76cf-40bd-8368-f52b1f74e841" containerID="b78d7e227a922b6e7062d987f0de9fb1db5488d06576147aac1094671774a1af" exitCode=0 Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.899731 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerDied","Data":"b78d7e227a922b6e7062d987f0de9fb1db5488d06576147aac1094671774a1af"} Mar 14 09:12:19 crc kubenswrapper[4843]: I0314 09:12:19.906893 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} Mar 14 09:12:20 crc kubenswrapper[4843]: I0314 09:12:20.914435 4843 generic.go:334] "Generic (PLEG): container finished" podID="da3d426b-76cf-40bd-8368-f52b1f74e841" containerID="3e0fbd239922d37b025a75c7c29aee3d3d0d6f20df354253d9e7c0c66c8d06cb" exitCode=0 Mar 14 09:12:20 crc kubenswrapper[4843]: I0314 09:12:20.914498 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerDied","Data":"3e0fbd239922d37b025a75c7c29aee3d3d0d6f20df354253d9e7c0c66c8d06cb"} Mar 14 09:12:21 crc kubenswrapper[4843]: I0314 09:12:21.338125 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:21 crc kubenswrapper[4843]: I0314 09:12:21.338165 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:21 crc kubenswrapper[4843]: E0314 09:12:21.338341 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:21 crc kubenswrapper[4843]: E0314 09:12:21.338451 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:21 crc kubenswrapper[4843]: I0314 09:12:21.338688 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:21 crc kubenswrapper[4843]: I0314 09:12:21.338876 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:21 crc kubenswrapper[4843]: E0314 09:12:21.338977 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:21 crc kubenswrapper[4843]: E0314 09:12:21.339068 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:21 crc kubenswrapper[4843]: I0314 09:12:21.936554 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerStarted","Data":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} Mar 14 09:12:21 crc kubenswrapper[4843]: I0314 09:12:21.943096 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerStarted","Data":"259f67dcaf16ac4f5b521b8fd70fc66337c7a6d31e210462fdf4cabb8cfe7ac0"} Mar 14 09:12:22 crc kubenswrapper[4843]: I0314 09:12:22.953476 4843 generic.go:334] "Generic (PLEG): container finished" podID="da3d426b-76cf-40bd-8368-f52b1f74e841" containerID="259f67dcaf16ac4f5b521b8fd70fc66337c7a6d31e210462fdf4cabb8cfe7ac0" exitCode=0 Mar 14 09:12:22 crc kubenswrapper[4843]: I0314 09:12:22.953594 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerDied","Data":"259f67dcaf16ac4f5b521b8fd70fc66337c7a6d31e210462fdf4cabb8cfe7ac0"} Mar 14 09:12:22 crc kubenswrapper[4843]: I0314 09:12:22.954073 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:22 crc kubenswrapper[4843]: I0314 09:12:22.954141 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:22 crc kubenswrapper[4843]: I0314 09:12:22.954802 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.004354 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.004820 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.036625 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podStartSLOduration=36.036591197 podStartE2EDuration="36.036591197s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:23.036103912 +0000 UTC m=+110.348715050" watchObservedRunningTime="2026-03-14 09:12:23.036591197 +0000 UTC m=+110.349202335" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.107777 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.107822 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.107927 4843 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.107947 4843 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.107987 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:39.107973519 +0000 UTC m=+126.420584647 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.108016 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:39.10799814 +0000 UTC m=+126.420609268 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.209188 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.209353 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209385 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:39.209357766 +0000 UTC m=+126.521968894 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209473 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209493 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209507 4843 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.209510 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209560 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:39.209542811 +0000 UTC m=+126.522153949 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209653 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209674 4843 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209688 4843 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.209733 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:39.209721146 +0000 UTC m=+126.522332284 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.310861 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.310981 4843 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.311047 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs podName:ca968925-6d0f-46b7-b12b-7c617faabd3c nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.311028481 +0000 UTC m=+118.623639629 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs") pod "network-metrics-daemon-2zndz" (UID: "ca968925-6d0f-46b7-b12b-7c617faabd3c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.339491 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.339559 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.339728 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.339574 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.339855 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.339953 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.340101 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:23 crc kubenswrapper[4843]: E0314 09:12:23.340171 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.961822 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" event={"ID":"da3d426b-76cf-40bd-8368-f52b1f74e841","Type":"ContainerStarted","Data":"9e84d3ed121f015aefdca26463287d9d0f318c4bc2f99f346b8c705d4a2af1c4"} Mar 14 09:12:23 crc kubenswrapper[4843]: I0314 09:12:23.980872 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-p8t7w" podStartSLOduration=37.980849305 podStartE2EDuration="37.980849305s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:23.980300459 +0000 UTC m=+111.292911597" watchObservedRunningTime="2026-03-14 09:12:23.980849305 +0000 UTC m=+111.293460443" Mar 14 09:12:24 crc kubenswrapper[4843]: I0314 09:12:24.290688 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2zndz"] Mar 14 09:12:24 crc kubenswrapper[4843]: I0314 09:12:24.290826 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:24 crc kubenswrapper[4843]: E0314 09:12:24.290931 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:25 crc kubenswrapper[4843]: I0314 09:12:25.338659 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:25 crc kubenswrapper[4843]: I0314 09:12:25.338695 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:25 crc kubenswrapper[4843]: E0314 09:12:25.339421 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:25 crc kubenswrapper[4843]: E0314 09:12:25.339594 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:25 crc kubenswrapper[4843]: I0314 09:12:25.338791 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:25 crc kubenswrapper[4843]: E0314 09:12:25.339760 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:26 crc kubenswrapper[4843]: I0314 09:12:26.338533 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:26 crc kubenswrapper[4843]: E0314 09:12:26.338747 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2zndz" podUID="ca968925-6d0f-46b7-b12b-7c617faabd3c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.339000 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.339007 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:27 crc kubenswrapper[4843]: E0314 09:12:27.339163 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.339129 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:27 crc kubenswrapper[4843]: E0314 09:12:27.339340 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 14 09:12:27 crc kubenswrapper[4843]: E0314 09:12:27.339664 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.416966 4843 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.417171 4843 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.478222 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-tr9zk"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.479001 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.487755 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.490049 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.500074 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.500373 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zkgfh"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.500874 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.501164 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.501207 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.502861 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mt9q4"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503171 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2j8wl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503358 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ghbdz"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.502870 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503470 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503371 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503068 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503122 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.503339 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.504098 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.508449 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-pwxsw"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.508894 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ph27v"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.522113 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.536193 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.536323 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-f2fl7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.536878 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.541264 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.541464 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.541640 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.541858 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542047 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542186 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542317 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542403 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542552 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542666 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.542943 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.543509 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.543692 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.543866 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.543964 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.544041 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.544114 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.544201 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.544259 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.544432 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.544620 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.549473 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.549818 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.549992 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.550793 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.550966 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.551447 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.551535 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.551682 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.551868 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.552057 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.552096 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.555223 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.555378 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.555683 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.555883 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.556086 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.556141 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.556498 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.556657 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.556792 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.556839 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.557061 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.557253 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.557436 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.558469 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.562875 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.562880 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-metrics-certs\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563416 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-client\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563470 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-config\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563498 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563556 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-trusted-ca\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563587 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb7125f9-9cca-41e8-9cb0-af829e1b7287-service-ca-bundle\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563610 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4aba2d14-caf3-485c-8ee7-725ed508d1cf-metrics-tls\") pod \"dns-operator-744455d44c-zkgfh\" (UID: \"4aba2d14-caf3-485c-8ee7-725ed508d1cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563554 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563661 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7x8s\" (UniqueName: \"kubernetes.io/projected/47b9849c-fc22-472c-a1f6-bd0e7a962a75-kube-api-access-d7x8s\") pod \"cluster-samples-operator-665b6dd947-vkqb9\" (UID: \"47b9849c-fc22-472c-a1f6-bd0e7a962a75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563738 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xndf5\" (UniqueName: \"kubernetes.io/projected/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-kube-api-access-xndf5\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563789 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-serving-cert\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563810 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-config\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563839 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9469\" (UniqueName: \"kubernetes.io/projected/1074b905-740e-4fc3-919b-e20309a9098b-kube-api-access-p9469\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563867 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1074b905-740e-4fc3-919b-e20309a9098b-serving-cert\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563893 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-service-ca\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563917 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/20cc2773-132c-4a74-9691-748a73111124-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563970 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggfst\" (UniqueName: \"kubernetes.io/projected/20cc2773-132c-4a74-9691-748a73111124-kube-api-access-ggfst\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563991 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq8nn\" (UniqueName: \"kubernetes.io/projected/4aba2d14-caf3-485c-8ee7-725ed508d1cf-kube-api-access-rq8nn\") pod \"dns-operator-744455d44c-zkgfh\" (UID: \"4aba2d14-caf3-485c-8ee7-725ed508d1cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564016 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-config\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564037 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/fb7125f9-9cca-41e8-9cb0-af829e1b7287-kube-api-access-txpj5\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564069 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prpr9\" (UniqueName: \"kubernetes.io/projected/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-kube-api-access-prpr9\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564092 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-oauth-serving-cert\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564117 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cc2773-132c-4a74-9691-748a73111124-serving-cert\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564144 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-default-certificate\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564170 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-serving-cert\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564189 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-service-ca\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564228 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-stats-auth\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564250 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-oauth-config\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564294 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-trusted-ca-bundle\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564318 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-ca\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564341 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljjp\" (UniqueName: \"kubernetes.io/projected/5c05075c-f721-4e78-ad5e-aaf88989336d-kube-api-access-5ljjp\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564374 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-client-ca\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564396 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-console-config\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564437 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/47b9849c-fc22-472c-a1f6-bd0e7a962a75-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vkqb9\" (UID: \"47b9849c-fc22-472c-a1f6-bd0e7a962a75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564477 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-serving-cert\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.564506 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvlb4\" (UniqueName: \"kubernetes.io/projected/570cf110-147b-482b-afe9-2b0ca133b6dc-kube-api-access-bvlb4\") pod \"downloads-7954f5f757-tr9zk\" (UID: \"570cf110-147b-482b-afe9-2b0ca133b6dc\") " pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.563621 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.565788 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.566566 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.566787 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.566972 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.567018 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.567109 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.567316 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.567365 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.570132 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.575562 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5kqvs"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.575718 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.576114 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.576549 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.592647 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.592818 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.593328 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.593777 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.595423 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.595936 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.598265 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xtq2f"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.602014 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.610438 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.610884 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n64hl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.611455 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.611481 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.611761 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.612065 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.612071 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.612162 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.612230 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.612684 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.612910 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.613341 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.613529 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.613709 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.613895 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.617085 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.618801 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.619317 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.620021 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.621096 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.621199 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.621533 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.621938 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.621967 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.622036 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.622225 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.622249 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.622433 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.622781 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.622918 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.623252 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.623558 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.623712 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.623851 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.623989 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.624156 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.624378 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.624796 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.624977 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625190 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625199 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625399 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625417 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625554 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625578 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625659 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hfz6k"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.625753 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.626019 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.626113 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.626175 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bfc6h"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.626492 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.627108 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.627251 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.627439 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.627469 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.627598 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.627674 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.628747 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.629631 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.630647 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.631938 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.632111 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zldnt"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.632681 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.632924 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.633714 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.633912 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.634077 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.634264 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.634460 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.634741 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.644345 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.645484 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jj2ff"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.646237 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.647689 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.649964 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.650341 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.659799 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.660505 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.663595 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.663709 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nxkww"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668786 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkswq\" (UniqueName: \"kubernetes.io/projected/82f06b03-5a8e-4ef6-aac7-3c0b9326183e-kube-api-access-fkswq\") pod \"control-plane-machine-set-operator-78cbb6b69f-cxr6c\" (UID: \"82f06b03-5a8e-4ef6-aac7-3c0b9326183e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668827 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668849 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668876 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-config\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668899 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668948 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cnzn\" (UniqueName: \"kubernetes.io/projected/664b59b6-607d-4320-8802-0ce25a001245-kube-api-access-5cnzn\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668974 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-trusted-ca\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.668995 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb7125f9-9cca-41e8-9cb0-af829e1b7287-service-ca-bundle\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669019 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4aba2d14-caf3-485c-8ee7-725ed508d1cf-metrics-tls\") pod \"dns-operator-744455d44c-zkgfh\" (UID: \"4aba2d14-caf3-485c-8ee7-725ed508d1cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669046 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7x8s\" (UniqueName: \"kubernetes.io/projected/47b9849c-fc22-472c-a1f6-bd0e7a962a75-kube-api-access-d7x8s\") pod \"cluster-samples-operator-665b6dd947-vkqb9\" (UID: \"47b9849c-fc22-472c-a1f6-bd0e7a962a75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669068 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/664b59b6-607d-4320-8802-0ce25a001245-audit-dir\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669094 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xndf5\" (UniqueName: \"kubernetes.io/projected/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-kube-api-access-xndf5\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669124 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-serving-cert\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669146 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-config\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669170 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669193 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669218 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9469\" (UniqueName: \"kubernetes.io/projected/1074b905-740e-4fc3-919b-e20309a9098b-kube-api-access-p9469\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669241 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669283 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1074b905-740e-4fc3-919b-e20309a9098b-serving-cert\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669308 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-service-ca\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669331 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/20cc2773-132c-4a74-9691-748a73111124-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669355 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6pk2\" (UniqueName: \"kubernetes.io/projected/8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6-kube-api-access-t6pk2\") pod \"migrator-59844c95c7-qxscp\" (UID: \"8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669395 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq8nn\" (UniqueName: \"kubernetes.io/projected/4aba2d14-caf3-485c-8ee7-725ed508d1cf-kube-api-access-rq8nn\") pod \"dns-operator-744455d44c-zkgfh\" (UID: \"4aba2d14-caf3-485c-8ee7-725ed508d1cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669417 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggfst\" (UniqueName: \"kubernetes.io/projected/20cc2773-132c-4a74-9691-748a73111124-kube-api-access-ggfst\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669441 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669461 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669488 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbctp\" (UniqueName: \"kubernetes.io/projected/bac1c5d6-a775-483f-ad66-20143a6b5141-kube-api-access-zbctp\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669514 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-config\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669544 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/fb7125f9-9cca-41e8-9cb0-af829e1b7287-kube-api-access-txpj5\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669572 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/694288bd-c353-44f3-806b-5ce7c6a10486-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hc4vf\" (UID: \"694288bd-c353-44f3-806b-5ce7c6a10486\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669600 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prpr9\" (UniqueName: \"kubernetes.io/projected/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-kube-api-access-prpr9\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669620 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-oauth-serving-cert\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669641 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cc2773-132c-4a74-9691-748a73111124-serving-cert\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669671 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-default-certificate\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669723 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-serving-cert\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669746 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-service-ca\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669767 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-trusted-ca-bundle\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669806 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-stats-auth\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669829 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-oauth-config\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669859 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-ca\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669899 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljjp\" (UniqueName: \"kubernetes.io/projected/5c05075c-f721-4e78-ad5e-aaf88989336d-kube-api-access-5ljjp\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669934 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-client-ca\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669955 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-console-config\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.669986 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670011 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670035 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670058 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/47b9849c-fc22-472c-a1f6-bd0e7a962a75-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vkqb9\" (UID: \"47b9849c-fc22-472c-a1f6-bd0e7a962a75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670078 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/82f06b03-5a8e-4ef6-aac7-3c0b9326183e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cxr6c\" (UID: \"82f06b03-5a8e-4ef6-aac7-3c0b9326183e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670092 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670100 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670125 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670149 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-serving-cert\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670172 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670198 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvlb4\" (UniqueName: \"kubernetes.io/projected/570cf110-147b-482b-afe9-2b0ca133b6dc-kube-api-access-bvlb4\") pod \"downloads-7954f5f757-tr9zk\" (UID: \"570cf110-147b-482b-afe9-2b0ca133b6dc\") " pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670220 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkrbj\" (UniqueName: \"kubernetes.io/projected/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-kube-api-access-fkrbj\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670242 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-audit-policies\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670283 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670308 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn7s6\" (UniqueName: \"kubernetes.io/projected/694288bd-c353-44f3-806b-5ce7c6a10486-kube-api-access-zn7s6\") pod \"package-server-manager-789f6589d5-hc4vf\" (UID: \"694288bd-c353-44f3-806b-5ce7c6a10486\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670331 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670353 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-client\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670376 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-metrics-certs\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.670399 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-proxy-tls\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.671969 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-config\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.673297 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.674231 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-trusted-ca\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.674439 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.675068 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/20cc2773-132c-4a74-9691-748a73111124-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.675394 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.675809 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-config\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.676725 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-client-ca\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.676771 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-oauth-serving-cert\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.677490 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-ca\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.677632 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-service-ca\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.677750 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-console-config\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.678079 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-config\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.678547 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb7125f9-9cca-41e8-9cb0-af829e1b7287-service-ca-bundle\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.679050 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.679477 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-trusted-ca-bundle\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.680140 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.680636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-service-ca\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.681954 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4aba2d14-caf3-485c-8ee7-725ed508d1cf-metrics-tls\") pod \"dns-operator-744455d44c-zkgfh\" (UID: \"4aba2d14-caf3-485c-8ee7-725ed508d1cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.682440 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-serving-cert\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.684729 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.684920 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1074b905-740e-4fc3-919b-e20309a9098b-serving-cert\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.684967 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tr9zk"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.685479 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-serving-cert\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.686072 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zkgfh"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.686591 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20cc2773-132c-4a74-9691-748a73111124-serving-cert\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.686948 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-oauth-config\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.687112 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-default-certificate\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.687209 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-serving-cert\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.687557 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-metrics-certs\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.687688 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fb7125f9-9cca-41e8-9cb0-af829e1b7287-stats-auth\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.687982 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-etcd-client\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.688592 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ghbdz"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.689049 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/47b9849c-fc22-472c-a1f6-bd0e7a962a75-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vkqb9\" (UID: \"47b9849c-fc22-472c-a1f6-bd0e7a962a75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.691941 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-swrz2"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.693208 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2j8wl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.693230 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-46z5v"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.693670 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.694120 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.694228 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.695466 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bfc6h"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.696602 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.697593 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.698547 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.699866 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n64hl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.701506 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.701906 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mt9q4"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.702935 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.703844 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.704852 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.706367 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-f2fl7"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.707535 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.709198 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xtq2f"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.710078 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5kqvs"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.710869 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.714183 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ph27v"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.714318 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mrmxq"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.715864 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xv469"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.716224 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.716358 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.716650 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.717314 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.720062 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.722810 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.724996 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.725673 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.729245 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.739053 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zldnt"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.741070 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.742041 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.743017 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.743991 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.745076 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hfz6k"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.746247 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jj2ff"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.747489 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.749074 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.750282 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nxkww"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.751189 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-46z5v"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.752421 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.753522 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-swrz2"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.754477 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-mr75n"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.755997 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mr75n"] Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.756111 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.762428 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.771816 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.771860 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.771888 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.771915 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/82f06b03-5a8e-4ef6-aac7-3c0b9326183e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cxr6c\" (UID: \"82f06b03-5a8e-4ef6-aac7-3c0b9326183e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.771951 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.771978 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772002 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772032 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkrbj\" (UniqueName: \"kubernetes.io/projected/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-kube-api-access-fkrbj\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772057 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-audit-policies\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772084 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772108 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn7s6\" (UniqueName: \"kubernetes.io/projected/694288bd-c353-44f3-806b-5ce7c6a10486-kube-api-access-zn7s6\") pod \"package-server-manager-789f6589d5-hc4vf\" (UID: \"694288bd-c353-44f3-806b-5ce7c6a10486\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772133 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772174 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-proxy-tls\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772203 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkswq\" (UniqueName: \"kubernetes.io/projected/82f06b03-5a8e-4ef6-aac7-3c0b9326183e-kube-api-access-fkswq\") pod \"control-plane-machine-set-operator-78cbb6b69f-cxr6c\" (UID: \"82f06b03-5a8e-4ef6-aac7-3c0b9326183e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772226 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772252 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772291 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cnzn\" (UniqueName: \"kubernetes.io/projected/664b59b6-607d-4320-8802-0ce25a001245-kube-api-access-5cnzn\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772333 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/664b59b6-607d-4320-8802-0ce25a001245-audit-dir\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772404 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772430 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772458 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772486 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6pk2\" (UniqueName: \"kubernetes.io/projected/8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6-kube-api-access-t6pk2\") pod \"migrator-59844c95c7-qxscp\" (UID: \"8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772540 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772563 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772482 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/664b59b6-607d-4320-8802-0ce25a001245-audit-dir\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772590 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbctp\" (UniqueName: \"kubernetes.io/projected/bac1c5d6-a775-483f-ad66-20143a6b5141-kube-api-access-zbctp\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.772638 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/694288bd-c353-44f3-806b-5ce7c6a10486-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hc4vf\" (UID: \"694288bd-c353-44f3-806b-5ce7c6a10486\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.773435 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.774722 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/82f06b03-5a8e-4ef6-aac7-3c0b9326183e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cxr6c\" (UID: \"82f06b03-5a8e-4ef6-aac7-3c0b9326183e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.782079 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.801323 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.821404 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.841559 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.861722 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.881924 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.885815 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/694288bd-c353-44f3-806b-5ce7c6a10486-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hc4vf\" (UID: \"694288bd-c353-44f3-806b-5ce7c6a10486\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.902156 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.921819 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.961080 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 14 09:12:27 crc kubenswrapper[4843]: I0314 09:12:27.980978 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.002155 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.021593 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.041215 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.061188 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.081954 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.101324 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.121996 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.142026 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.161320 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.181952 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.201373 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.221920 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.241730 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.261818 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.281749 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.301509 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.321582 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.327002 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-proxy-tls\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.339406 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.342728 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.373968 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.380986 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.388816 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.401991 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.406915 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.421618 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.427054 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.442039 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.467687 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.478039 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.481382 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.486569 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.501796 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.521346 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.526582 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.541529 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.561316 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.583445 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.603341 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.622107 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.641601 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.646906 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.660117 4843 request.go:700] Waited for 1.001074401s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/configmaps?fieldSelector=metadata.name%3Daudit&limit=500&resourceVersion=0 Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.662009 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.663452 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-audit-policies\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.681998 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.701603 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.704128 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.722335 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.728896 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.742991 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.771534 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773245 4843 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773403 4843 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773342 4843 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773338 4843 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773615 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config podName:bac1c5d6-a775-483f-ad66-20143a6b5141 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:29.273421145 +0000 UTC m=+116.586032273 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config") pod "route-controller-manager-6576b87f9c-4nzm6" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141") : failed to sync configmap cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773739 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca podName:664b59b6-607d-4320-8802-0ce25a001245 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:29.273718113 +0000 UTC m=+116.586329241 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-jj2ff" (UID: "664b59b6-607d-4320-8802-0ce25a001245") : failed to sync configmap cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773806 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert podName:bac1c5d6-a775-483f-ad66-20143a6b5141 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:29.273798716 +0000 UTC m=+116.586409844 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert") pod "route-controller-manager-6576b87f9c-4nzm6" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141") : failed to sync secret cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: E0314 09:12:28.773867 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca podName:bac1c5d6-a775-483f-ad66-20143a6b5141 nodeName:}" failed. No retries permitted until 2026-03-14 09:12:29.273860258 +0000 UTC m=+116.586471386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca") pod "route-controller-manager-6576b87f9c-4nzm6" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141") : failed to sync configmap cache: timed out waiting for the condition Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.774673 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.781854 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.801686 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.821111 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.842411 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.881936 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.901755 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.921725 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.945222 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.962588 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 14 09:12:28 crc kubenswrapper[4843]: I0314 09:12:28.981633 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.012400 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.021147 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.058222 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvlb4\" (UniqueName: \"kubernetes.io/projected/570cf110-147b-482b-afe9-2b0ca133b6dc-kube-api-access-bvlb4\") pod \"downloads-7954f5f757-tr9zk\" (UID: \"570cf110-147b-482b-afe9-2b0ca133b6dc\") " pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.081122 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq8nn\" (UniqueName: \"kubernetes.io/projected/4aba2d14-caf3-485c-8ee7-725ed508d1cf-kube-api-access-rq8nn\") pod \"dns-operator-744455d44c-zkgfh\" (UID: \"4aba2d14-caf3-485c-8ee7-725ed508d1cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.100888 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.101500 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggfst\" (UniqueName: \"kubernetes.io/projected/20cc2773-132c-4a74-9691-748a73111124-kube-api-access-ggfst\") pod \"openshift-config-operator-7777fb866f-ph27v\" (UID: \"20cc2773-132c-4a74-9691-748a73111124\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.121174 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.142258 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.147558 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.160879 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.181508 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.202088 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.222698 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.262708 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.270867 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txpj5\" (UniqueName: \"kubernetes.io/projected/fb7125f9-9cca-41e8-9cb0-af829e1b7287-kube-api-access-txpj5\") pod \"router-default-5444994796-pwxsw\" (UID: \"fb7125f9-9cca-41e8-9cb0-af829e1b7287\") " pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.299594 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.299675 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.299718 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.299787 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.300551 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.301143 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.301781 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.303767 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.339388 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.339598 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.339818 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.340104 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.341190 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.346083 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljjp\" (UniqueName: \"kubernetes.io/projected/5c05075c-f721-4e78-ad5e-aaf88989336d-kube-api-access-5ljjp\") pod \"console-f9d7485db-2j8wl\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.355548 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9469\" (UniqueName: \"kubernetes.io/projected/1074b905-740e-4fc3-919b-e20309a9098b-kube-api-access-p9469\") pod \"controller-manager-879f6c89f-mt9q4\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.372242 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7x8s\" (UniqueName: \"kubernetes.io/projected/47b9849c-fc22-472c-a1f6-bd0e7a962a75-kube-api-access-d7x8s\") pod \"cluster-samples-operator-665b6dd947-vkqb9\" (UID: \"47b9849c-fc22-472c-a1f6-bd0e7a962a75\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.376588 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.381132 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.383496 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prpr9\" (UniqueName: \"kubernetes.io/projected/a46e0a56-02cc-43c0-b7ab-1c792b04fa2b-kube-api-access-prpr9\") pod \"console-operator-58897d9998-ghbdz\" (UID: \"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b\") " pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.400297 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.402236 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.420018 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.422251 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.431568 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.441006 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ph27v"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.441838 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 14 09:12:29 crc kubenswrapper[4843]: W0314 09:12:29.451901 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20cc2773_132c_4a74_9691_748a73111124.slice/crio-84b6988762677591d13026568fecee63237ab85d75b7aab0f6554bb77c84cd2c WatchSource:0}: Error finding container 84b6988762677591d13026568fecee63237ab85d75b7aab0f6554bb77c84cd2c: Status 404 returned error can't find the container with id 84b6988762677591d13026568fecee63237ab85d75b7aab0f6554bb77c84cd2c Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.462903 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.483811 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.521654 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.523390 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xndf5\" (UniqueName: \"kubernetes.io/projected/be6a1160-25cd-4a04-ad30-f1c40c8d7e05-kube-api-access-xndf5\") pod \"etcd-operator-b45778765-f2fl7\" (UID: \"be6a1160-25cd-4a04-ad30-f1c40c8d7e05\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.535445 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.535639 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.540239 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tr9zk"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.542578 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.561295 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.581183 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.597642 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zkgfh"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.602006 4843 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 14 09:12:29 crc kubenswrapper[4843]: W0314 09:12:29.608687 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aba2d14_caf3_485c_8ee7_725ed508d1cf.slice/crio-719d4077ffb3a4f07f19b1d5becddea7234537bebbe5e6f787d9053f52de46ed WatchSource:0}: Error finding container 719d4077ffb3a4f07f19b1d5becddea7234537bebbe5e6f787d9053f52de46ed: Status 404 returned error can't find the container with id 719d4077ffb3a4f07f19b1d5becddea7234537bebbe5e6f787d9053f52de46ed Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.621591 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.641246 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.661165 4843 request.go:700] Waited for 1.944502175s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.663021 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ghbdz"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.663322 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.671971 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.682355 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.702403 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.722399 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.741663 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.762050 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.788349 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.813368 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-f2fl7"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.829933 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkrbj\" (UniqueName: \"kubernetes.io/projected/fb768ff1-000e-4bef-9c37-fcfb4fdd7e76-kube-api-access-fkrbj\") pod \"machine-config-controller-84d6567774-xvzgq\" (UID: \"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.834741 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2j8wl"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.836047 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mt9q4"] Mar 14 09:12:29 crc kubenswrapper[4843]: W0314 09:12:29.847082 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c05075c_f721_4e78_ad5e_aaf88989336d.slice/crio-9766f82fbbc308f68ee4c376c9e33a6cb44cc45b9a50df672f215295a9ad922d WatchSource:0}: Error finding container 9766f82fbbc308f68ee4c376c9e33a6cb44cc45b9a50df672f215295a9ad922d: Status 404 returned error can't find the container with id 9766f82fbbc308f68ee4c376c9e33a6cb44cc45b9a50df672f215295a9ad922d Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.847836 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cnzn\" (UniqueName: \"kubernetes.io/projected/664b59b6-607d-4320-8802-0ce25a001245-kube-api-access-5cnzn\") pod \"oauth-openshift-558db77b4-jj2ff\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.860437 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn7s6\" (UniqueName: \"kubernetes.io/projected/694288bd-c353-44f3-806b-5ce7c6a10486-kube-api-access-zn7s6\") pod \"package-server-manager-789f6589d5-hc4vf\" (UID: \"694288bd-c353-44f3-806b-5ce7c6a10486\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:29 crc kubenswrapper[4843]: W0314 09:12:29.861621 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1074b905_740e_4fc3_919b_e20309a9098b.slice/crio-5f7c4a1b7081a880f02daf9170c25ef705ca33472782d639502359e9c32e769e WatchSource:0}: Error finding container 5f7c4a1b7081a880f02daf9170c25ef705ca33472782d639502359e9c32e769e: Status 404 returned error can't find the container with id 5f7c4a1b7081a880f02daf9170c25ef705ca33472782d639502359e9c32e769e Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.877804 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkswq\" (UniqueName: \"kubernetes.io/projected/82f06b03-5a8e-4ef6-aac7-3c0b9326183e-kube-api-access-fkswq\") pod \"control-plane-machine-set-operator-78cbb6b69f-cxr6c\" (UID: \"82f06b03-5a8e-4ef6-aac7-3c0b9326183e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.884186 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9"] Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.900561 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6pk2\" (UniqueName: \"kubernetes.io/projected/8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6-kube-api-access-t6pk2\") pod \"migrator-59844c95c7-qxscp\" (UID: \"8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.907942 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.917156 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbctp\" (UniqueName: \"kubernetes.io/projected/bac1c5d6-a775-483f-ad66-20143a6b5141-kube-api-access-zbctp\") pod \"route-controller-manager-6576b87f9c-4nzm6\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.921097 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.941570 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.949047 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.968786 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.986172 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" event={"ID":"be6a1160-25cd-4a04-ad30-f1c40c8d7e05","Type":"ContainerStarted","Data":"b26e13ccb490745fd31310828b5525c690a7709000b3a771c6b4e6de1d31876e"} Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.989857 4843 generic.go:334] "Generic (PLEG): container finished" podID="20cc2773-132c-4a74-9691-748a73111124" containerID="16b4eb900e12be1e92f5b20401248cdf237d07810085a42bdf6d38cf579ccb56" exitCode=0 Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.989939 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" event={"ID":"20cc2773-132c-4a74-9691-748a73111124","Type":"ContainerDied","Data":"16b4eb900e12be1e92f5b20401248cdf237d07810085a42bdf6d38cf579ccb56"} Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.989979 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" event={"ID":"20cc2773-132c-4a74-9691-748a73111124","Type":"ContainerStarted","Data":"84b6988762677591d13026568fecee63237ab85d75b7aab0f6554bb77c84cd2c"} Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.993558 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" event={"ID":"47b9849c-fc22-472c-a1f6-bd0e7a962a75","Type":"ContainerStarted","Data":"a5f6d41788f951126eeb25279ffe513e191d7fd95c6f756aeac42415c75a77ce"} Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.997123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" event={"ID":"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b","Type":"ContainerStarted","Data":"1cb37b8a243b87ac066db2eafd8798f7cc40b94f404a5bdcbe84960ef04c604f"} Mar 14 09:12:29 crc kubenswrapper[4843]: I0314 09:12:29.997424 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.001220 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.002349 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" event={"ID":"4aba2d14-caf3-485c-8ee7-725ed508d1cf","Type":"ContainerStarted","Data":"ab6e6550bf6096adb6408ddd332f60c379b7fc3668f73265eff2a5d688c9e15a"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.002747 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" event={"ID":"4aba2d14-caf3-485c-8ee7-725ed508d1cf","Type":"ContainerStarted","Data":"719d4077ffb3a4f07f19b1d5becddea7234537bebbe5e6f787d9053f52de46ed"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.004009 4843 patch_prober.go:28] interesting pod/console-operator-58897d9998-ghbdz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.004088 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" podUID="a46e0a56-02cc-43c0-b7ab-1c792b04fa2b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010290 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e221699-471c-4cec-9aa8-b680923ac8b9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010320 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30f8e198-f784-4504-8fa3-09b59daa7b4d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010337 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgpzp\" (UniqueName: \"kubernetes.io/projected/e21032b1-6e5a-480c-92df-d60d83844fc2-kube-api-access-vgpzp\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010355 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj4wd\" (UniqueName: \"kubernetes.io/projected/e037c68c-36ae-47b2-b3f5-e159c512cf72-kube-api-access-cj4wd\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010382 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds9h2\" (UniqueName: \"kubernetes.io/projected/9fafc3ce-d8b1-4179-be51-d7d3f0c17217-kube-api-access-ds9h2\") pod \"multus-admission-controller-857f4d67dd-bfc6h\" (UID: \"9fafc3ce-d8b1-4179-be51-d7d3f0c17217\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010461 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e21032b1-6e5a-480c-92df-d60d83844fc2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010477 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e221699-471c-4cec-9aa8-b680923ac8b9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010496 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg86s\" (UniqueName: \"kubernetes.io/projected/66202f9d-911e-47f7-b2d2-6326dd9c7283-kube-api-access-cg86s\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010510 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dd09639-861b-4f3d-b001-e28c8650cad7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010526 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1c2692c-98ff-4bd0-8c81-8249f124b464-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010541 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e21032b1-6e5a-480c-92df-d60d83844fc2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010560 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mln7r\" (UniqueName: \"kubernetes.io/projected/e494d22d-e0ff-4fab-85d6-5241d72ff039-kube-api-access-mln7r\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010588 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6fc3d87-f84d-4782-8cab-adf072d642b4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010620 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010636 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-certificates\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010653 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce314766-eb86-4b0a-909b-2ae24850a4af-config\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010679 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010711 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30f8e198-f784-4504-8fa3-09b59daa7b4d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010725 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-image-import-ca\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010793 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g77xg\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-kube-api-access-g77xg\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010808 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010827 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnfxk\" (UniqueName: \"kubernetes.io/projected/c4b10c60-ed92-4bcf-9385-bf169e81051f-kube-api-access-nnfxk\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010846 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66202f9d-911e-47f7-b2d2-6326dd9c7283-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010863 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swrkm\" (UniqueName: \"kubernetes.io/projected/35935053-e284-4215-b14c-efcb387201c4-kube-api-access-swrkm\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010877 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dcf6a857-49d2-4319-b083-3ff05a19191b-signing-key\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010892 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-audit\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010906 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dd09639-861b-4f3d-b001-e28c8650cad7-config\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010922 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15553d9e-d0a1-495b-8c4f-35f3d026fb33-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010937 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-node-pullsecrets\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.010955 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e037c68c-36ae-47b2-b3f5-e159c512cf72-serving-cert\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011150 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011182 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15553d9e-d0a1-495b-8c4f-35f3d026fb33-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011197 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9fafc3ce-d8b1-4179-be51-d7d3f0c17217-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bfc6h\" (UID: \"9fafc3ce-d8b1-4179-be51-d7d3f0c17217\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011214 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-tls\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011228 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-config\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011243 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wzqt\" (UniqueName: \"kubernetes.io/projected/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-kube-api-access-7wzqt\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011258 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-tmpfs\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011322 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35935053-e284-4215-b14c-efcb387201c4-secret-volume\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011352 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tgq2\" (UniqueName: \"kubernetes.io/projected/30f8e198-f784-4504-8fa3-09b59daa7b4d-kube-api-access-9tgq2\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011366 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011381 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-serving-cert\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011398 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce314766-eb86-4b0a-909b-2ae24850a4af-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011412 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66202f9d-911e-47f7-b2d2-6326dd9c7283-proxy-tls\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011429 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011443 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4b10c60-ed92-4bcf-9385-bf169e81051f-srv-cert\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011457 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4b10c60-ed92-4bcf-9385-bf169e81051f-profile-collector-cert\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011499 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-trusted-ca\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011538 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dcf6a857-49d2-4319-b083-3ff05a19191b-signing-cabundle\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011585 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-bound-sa-token\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011619 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e494d22d-e0ff-4fab-85d6-5241d72ff039-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011637 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mvq8\" (UniqueName: \"kubernetes.io/projected/f1c2692c-98ff-4bd0-8c81-8249f124b464-kube-api-access-6mvq8\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011652 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15553d9e-d0a1-495b-8c4f-35f3d026fb33-config\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011667 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e21032b1-6e5a-480c-92df-d60d83844fc2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011684 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6s4x\" (UniqueName: \"kubernetes.io/projected/dcf6a857-49d2-4319-b083-3ff05a19191b-kube-api-access-l6s4x\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011710 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1c2692c-98ff-4bd0-8c81-8249f124b464-trusted-ca\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011750 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-audit-dir\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011767 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e037c68c-36ae-47b2-b3f5-e159c512cf72-config\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011798 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6fc3d87-f84d-4782-8cab-adf072d642b4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011815 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011830 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-encryption-config\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011845 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011861 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pnk2\" (UniqueName: \"kubernetes.io/projected/991014d6-3d2c-4291-91ef-db919b1c7d1c-kube-api-access-9pnk2\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011879 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35935053-e284-4215-b14c-efcb387201c4-config-volume\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011895 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72fbw\" (UniqueName: \"kubernetes.io/projected/6e221699-471c-4cec-9aa8-b680923ac8b9-kube-api-access-72fbw\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011912 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9pwz\" (UniqueName: \"kubernetes.io/projected/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-kube-api-access-c9pwz\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011935 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e494d22d-e0ff-4fab-85d6-5241d72ff039-srv-cert\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011951 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-webhook-cert\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011964 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011978 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-etcd-client\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.011994 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66202f9d-911e-47f7-b2d2-6326dd9c7283-images\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012009 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dd09639-861b-4f3d-b001-e28c8650cad7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012024 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-apiservice-cert\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012039 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1c2692c-98ff-4bd0-8c81-8249f124b464-metrics-tls\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012055 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhgmq\" (UniqueName: \"kubernetes.io/projected/e6fc3d87-f84d-4782-8cab-adf072d642b4-kube-api-access-bhgmq\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012071 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jjfc\" (UniqueName: \"kubernetes.io/projected/ce314766-eb86-4b0a-909b-2ae24850a4af-kube-api-access-2jjfc\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012101 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.012119 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ce314766-eb86-4b0a-909b-2ae24850a4af-images\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.019869 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:30.519853343 +0000 UTC m=+117.832464471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.020101 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.021370 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.027353 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2j8wl" event={"ID":"5c05075c-f721-4e78-ad5e-aaf88989336d","Type":"ContainerStarted","Data":"9766f82fbbc308f68ee4c376c9e33a6cb44cc45b9a50df672f215295a9ad922d"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.029131 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.031392 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" event={"ID":"1074b905-740e-4fc3-919b-e20309a9098b","Type":"ContainerStarted","Data":"5f7c4a1b7081a880f02daf9170c25ef705ca33472782d639502359e9c32e769e"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.043185 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.045904 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.051303 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pwxsw" event={"ID":"fb7125f9-9cca-41e8-9cb0-af829e1b7287","Type":"ContainerStarted","Data":"d3615eb0a29fe6b72150e25a27a6da7264d12467713b38d580d80392c5154ae6"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.051349 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pwxsw" event={"ID":"fb7125f9-9cca-41e8-9cb0-af829e1b7287","Type":"ContainerStarted","Data":"e5134cca46a5167103aefc22dd14cd72e89ba8653178b1883c69486e2fae3844"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.054001 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tr9zk" event={"ID":"570cf110-147b-482b-afe9-2b0ca133b6dc","Type":"ContainerStarted","Data":"194ae86bab6d61be76cd41467a9b309b42214adbe49d679d354115e34a2544f6"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.054045 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tr9zk" event={"ID":"570cf110-147b-482b-afe9-2b0ca133b6dc","Type":"ContainerStarted","Data":"98fc668ffcc4214f005fa587c28bae23051c040d591fe565c4777f9cf941d57c"} Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.054705 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.059364 4843 patch_prober.go:28] interesting pod/downloads-7954f5f757-tr9zk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.059948 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tr9zk" podUID="570cf110-147b-482b-afe9-2b0ca133b6dc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.061008 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.113805 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114030 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9fafc3ce-d8b1-4179-be51-d7d3f0c17217-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bfc6h\" (UID: \"9fafc3ce-d8b1-4179-be51-d7d3f0c17217\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114059 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15553d9e-d0a1-495b-8c4f-35f3d026fb33-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114088 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlwq6\" (UniqueName: \"kubernetes.io/projected/6cd2a062-18f8-44d1-918e-02142c6a2c16-kube-api-access-tlwq6\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114115 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/94325ed6-8249-4f3a-98fc-fc507ab5ec18-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114167 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncrfv\" (UniqueName: \"kubernetes.io/projected/26b0b618-116b-4aa2-8e84-3e024a2c46fd-kube-api-access-ncrfv\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114199 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-tls\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114215 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-config\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114231 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wzqt\" (UniqueName: \"kubernetes.io/projected/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-kube-api-access-7wzqt\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114248 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-tmpfs\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114284 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-certs\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114345 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35935053-e284-4215-b14c-efcb387201c4-secret-volume\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114366 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cd2a062-18f8-44d1-918e-02142c6a2c16-config-volume\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114394 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tgq2\" (UniqueName: \"kubernetes.io/projected/30f8e198-f784-4504-8fa3-09b59daa7b4d-kube-api-access-9tgq2\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114410 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114426 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-serving-cert\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114443 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce314766-eb86-4b0a-909b-2ae24850a4af-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114463 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66202f9d-911e-47f7-b2d2-6326dd9c7283-proxy-tls\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114479 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6rsb\" (UniqueName: \"kubernetes.io/projected/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-kube-api-access-w6rsb\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114498 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-registration-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114517 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114533 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4b10c60-ed92-4bcf-9385-bf169e81051f-srv-cert\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114549 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4b10c60-ed92-4bcf-9385-bf169e81051f-profile-collector-cert\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.114666 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:30.614588689 +0000 UTC m=+117.927199847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114749 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114803 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d978f007-fe7a-4956-afc5-c83101961ee8-serving-cert\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114830 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-mountpoint-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114873 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26b0b618-116b-4aa2-8e84-3e024a2c46fd-auth-proxy-config\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.114969 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-trusted-ca\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115019 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dcf6a857-49d2-4319-b083-3ff05a19191b-signing-cabundle\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115062 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-serving-cert\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115105 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6cd2a062-18f8-44d1-918e-02142c6a2c16-metrics-tls\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115128 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-service-ca-bundle\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115192 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-bound-sa-token\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115253 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e494d22d-e0ff-4fab-85d6-5241d72ff039-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115300 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26b0b618-116b-4aa2-8e84-3e024a2c46fd-machine-approver-tls\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115328 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mvq8\" (UniqueName: \"kubernetes.io/projected/f1c2692c-98ff-4bd0-8c81-8249f124b464-kube-api-access-6mvq8\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115369 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15553d9e-d0a1-495b-8c4f-35f3d026fb33-config\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115391 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e21032b1-6e5a-480c-92df-d60d83844fc2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115415 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6s4x\" (UniqueName: \"kubernetes.io/projected/dcf6a857-49d2-4319-b083-3ff05a19191b-kube-api-access-l6s4x\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115460 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4vl7\" (UniqueName: \"kubernetes.io/projected/d978f007-fe7a-4956-afc5-c83101961ee8-kube-api-access-n4vl7\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115483 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-node-bootstrap-token\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115538 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1c2692c-98ff-4bd0-8c81-8249f124b464-trusted-ca\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115560 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-socket-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115645 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-audit-dir\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115689 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e037c68c-36ae-47b2-b3f5-e159c512cf72-config\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115722 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-config\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115802 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6fc3d87-f84d-4782-8cab-adf072d642b4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115845 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72fbw\" (UniqueName: \"kubernetes.io/projected/6e221699-471c-4cec-9aa8-b680923ac8b9-kube-api-access-72fbw\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115900 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115953 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-encryption-config\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.115976 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116016 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pnk2\" (UniqueName: \"kubernetes.io/projected/991014d6-3d2c-4291-91ef-db919b1c7d1c-kube-api-access-9pnk2\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116037 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35935053-e284-4215-b14c-efcb387201c4-config-volume\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116115 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9pwz\" (UniqueName: \"kubernetes.io/projected/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-kube-api-access-c9pwz\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116144 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9gfq\" (UniqueName: \"kubernetes.io/projected/5921a039-7ee3-49ce-aad8-cf711e43e797-kube-api-access-f9gfq\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116206 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-webhook-cert\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116232 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e494d22d-e0ff-4fab-85d6-5241d72ff039-srv-cert\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116301 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116325 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-etcd-client\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116350 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66202f9d-911e-47f7-b2d2-6326dd9c7283-images\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.116398 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dd09639-861b-4f3d-b001-e28c8650cad7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.117317 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6fc3d87-f84d-4782-8cab-adf072d642b4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.118928 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.119350 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e21032b1-6e5a-480c-92df-d60d83844fc2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120069 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-csi-data-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120125 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/94325ed6-8249-4f3a-98fc-fc507ab5ec18-ready\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120199 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-apiservice-cert\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120225 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1c2692c-98ff-4bd0-8c81-8249f124b464-metrics-tls\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120329 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhgmq\" (UniqueName: \"kubernetes.io/projected/e6fc3d87-f84d-4782-8cab-adf072d642b4-kube-api-access-bhgmq\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120354 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jjfc\" (UniqueName: \"kubernetes.io/projected/ce314766-eb86-4b0a-909b-2ae24850a4af-kube-api-access-2jjfc\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120371 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8eefdff-74f7-4bb8-b39f-753a1375a403-cert\") pod \"ingress-canary-46z5v\" (UID: \"e8eefdff-74f7-4bb8-b39f-753a1375a403\") " pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120398 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120416 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ce314766-eb86-4b0a-909b-2ae24850a4af-images\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120481 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e221699-471c-4cec-9aa8-b680923ac8b9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120546 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30f8e198-f784-4504-8fa3-09b59daa7b4d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120567 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgpzp\" (UniqueName: \"kubernetes.io/projected/e21032b1-6e5a-480c-92df-d60d83844fc2-kube-api-access-vgpzp\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120584 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj4wd\" (UniqueName: \"kubernetes.io/projected/e037c68c-36ae-47b2-b3f5-e159c512cf72-kube-api-access-cj4wd\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120608 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-encryption-config\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.120635 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121311 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds9h2\" (UniqueName: \"kubernetes.io/projected/9fafc3ce-d8b1-4179-be51-d7d3f0c17217-kube-api-access-ds9h2\") pod \"multus-admission-controller-857f4d67dd-bfc6h\" (UID: \"9fafc3ce-d8b1-4179-be51-d7d3f0c17217\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121427 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e21032b1-6e5a-480c-92df-d60d83844fc2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121457 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-etcd-client\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121482 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg86s\" (UniqueName: \"kubernetes.io/projected/66202f9d-911e-47f7-b2d2-6326dd9c7283-kube-api-access-cg86s\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121505 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e221699-471c-4cec-9aa8-b680923ac8b9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121531 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b0b618-116b-4aa2-8e84-3e024a2c46fd-config\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121553 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9kfz\" (UniqueName: \"kubernetes.io/projected/94325ed6-8249-4f3a-98fc-fc507ab5ec18-kube-api-access-h9kfz\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121580 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dd09639-861b-4f3d-b001-e28c8650cad7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121625 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1c2692c-98ff-4bd0-8c81-8249f124b464-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121678 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e21032b1-6e5a-480c-92df-d60d83844fc2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121717 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mln7r\" (UniqueName: \"kubernetes.io/projected/e494d22d-e0ff-4fab-85d6-5241d72ff039-kube-api-access-mln7r\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121801 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6fc3d87-f84d-4782-8cab-adf072d642b4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121855 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121884 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-certificates\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121936 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce314766-eb86-4b0a-909b-2ae24850a4af-config\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.121965 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-plugins-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.122037 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.122080 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30f8e198-f784-4504-8fa3-09b59daa7b4d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.122106 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-image-import-ca\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.122162 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.135652 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15553d9e-d0a1-495b-8c4f-35f3d026fb33-config\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.136795 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66202f9d-911e-47f7-b2d2-6326dd9c7283-images\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.138163 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-etcd-client\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.138457 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-encryption-config\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.139000 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4b10c60-ed92-4bcf-9385-bf169e81051f-profile-collector-cert\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.139419 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66202f9d-911e-47f7-b2d2-6326dd9c7283-proxy-tls\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.139577 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dcf6a857-49d2-4319-b083-3ff05a19191b-signing-cabundle\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.139872 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-serving-cert\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.140424 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.141922 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e221699-471c-4cec-9aa8-b680923ac8b9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.142187 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9fafc3ce-d8b1-4179-be51-d7d3f0c17217-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bfc6h\" (UID: \"9fafc3ce-d8b1-4179-be51-d7d3f0c17217\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.142759 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-webhook-cert\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.144861 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-trusted-ca\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.145571 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:30.645519791 +0000 UTC m=+117.958130919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.146946 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ce314766-eb86-4b0a-909b-2ae24850a4af-images\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.147004 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-audit-dir\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.147730 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35935053-e284-4215-b14c-efcb387201c4-config-volume\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.148774 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15553d9e-d0a1-495b-8c4f-35f3d026fb33-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.149904 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e494d22d-e0ff-4fab-85d6-5241d72ff039-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.155003 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e037c68c-36ae-47b2-b3f5-e159c512cf72-config\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.155248 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.156066 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1c2692c-98ff-4bd0-8c81-8249f124b464-metrics-tls\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.157344 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35935053-e284-4215-b14c-efcb387201c4-secret-volume\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.157826 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4b10c60-ed92-4bcf-9385-bf169e81051f-srv-cert\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.161003 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-apiservice-cert\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.161243 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.163968 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce314766-eb86-4b0a-909b-2ae24850a4af-config\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.163980 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce314766-eb86-4b0a-909b-2ae24850a4af-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165083 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-config\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165734 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g77xg\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-kube-api-access-g77xg\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165770 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-audit-policies\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165794 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8glqh\" (UniqueName: \"kubernetes.io/projected/1724971d-d75c-4e8c-a951-0f10e9c5e799-kube-api-access-8glqh\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165818 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165965 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnfxk\" (UniqueName: \"kubernetes.io/projected/c4b10c60-ed92-4bcf-9385-bf169e81051f-kube-api-access-nnfxk\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.165995 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swrkm\" (UniqueName: \"kubernetes.io/projected/35935053-e284-4215-b14c-efcb387201c4-kube-api-access-swrkm\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166021 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-image-import-ca\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166036 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66202f9d-911e-47f7-b2d2-6326dd9c7283-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166062 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/94325ed6-8249-4f3a-98fc-fc507ab5ec18-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166089 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dcf6a857-49d2-4319-b083-3ff05a19191b-signing-key\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166146 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-audit\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166168 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dd09639-861b-4f3d-b001-e28c8650cad7-config\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166190 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15553d9e-d0a1-495b-8c4f-35f3d026fb33-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166214 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72ljv\" (UniqueName: \"kubernetes.io/projected/e8eefdff-74f7-4bb8-b39f-753a1375a403-kube-api-access-72ljv\") pod \"ingress-canary-46z5v\" (UID: \"e8eefdff-74f7-4bb8-b39f-753a1375a403\") " pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166250 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-node-pullsecrets\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166289 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e037c68c-36ae-47b2-b3f5-e159c512cf72-serving-cert\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166315 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-tmpfs\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166326 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.166347 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1724971d-d75c-4e8c-a951-0f10e9c5e799-audit-dir\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.167652 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1c2692c-98ff-4bd0-8c81-8249f124b464-trusted-ca\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.168084 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.168433 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dd09639-861b-4f3d-b001-e28c8650cad7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.168701 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e221699-471c-4cec-9aa8-b680923ac8b9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.169201 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66202f9d-911e-47f7-b2d2-6326dd9c7283-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.170987 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-audit\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.172327 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dd09639-861b-4f3d-b001-e28c8650cad7-config\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.174004 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.174004 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6fc3d87-f84d-4782-8cab-adf072d642b4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.174107 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30f8e198-f784-4504-8fa3-09b59daa7b4d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.174481 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-node-pullsecrets\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.174575 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.179724 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72fbw\" (UniqueName: \"kubernetes.io/projected/6e221699-471c-4cec-9aa8-b680923ac8b9-kube-api-access-72fbw\") pod \"kube-storage-version-migrator-operator-b67b599dd-kd9zf\" (UID: \"6e221699-471c-4cec-9aa8-b680923ac8b9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.183071 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e21032b1-6e5a-480c-92df-d60d83844fc2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.183363 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-tls\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.183570 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.183781 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30f8e198-f784-4504-8fa3-09b59daa7b4d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.184121 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dcf6a857-49d2-4319-b083-3ff05a19191b-signing-key\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.185085 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-certificates\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.191321 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e037c68c-36ae-47b2-b3f5-e159c512cf72-serving-cert\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.191800 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.191864 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e494d22d-e0ff-4fab-85d6-5241d72ff039-srv-cert\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.208657 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6s4x\" (UniqueName: \"kubernetes.io/projected/dcf6a857-49d2-4319-b083-3ff05a19191b-kube-api-access-l6s4x\") pod \"service-ca-9c57cc56f-zldnt\" (UID: \"dcf6a857-49d2-4319-b083-3ff05a19191b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.214289 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-bound-sa-token\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.238539 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9pwz\" (UniqueName: \"kubernetes.io/projected/8278bb86-cd39-4ecb-b0c6-75ec86c0763a-kube-api-access-c9pwz\") pod \"packageserver-d55dfcdfc-f5gt8\" (UID: \"8278bb86-cd39-4ecb-b0c6-75ec86c0763a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.251344 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhgmq\" (UniqueName: \"kubernetes.io/projected/e6fc3d87-f84d-4782-8cab-adf072d642b4-kube-api-access-bhgmq\") pod \"openshift-apiserver-operator-796bbdcf4f-xjk7w\" (UID: \"e6fc3d87-f84d-4782-8cab-adf072d642b4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.256339 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.258851 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jjfc\" (UniqueName: \"kubernetes.io/projected/ce314766-eb86-4b0a-909b-2ae24850a4af-kube-api-access-2jjfc\") pod \"machine-api-operator-5694c8668f-5kqvs\" (UID: \"ce314766-eb86-4b0a-909b-2ae24850a4af\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.259309 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267226 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267435 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8eefdff-74f7-4bb8-b39f-753a1375a403-cert\") pod \"ingress-canary-46z5v\" (UID: \"e8eefdff-74f7-4bb8-b39f-753a1375a403\") " pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267522 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-encryption-config\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267544 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267606 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-etcd-client\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267628 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b0b618-116b-4aa2-8e84-3e024a2c46fd-config\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.267663 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:30.767627113 +0000 UTC m=+118.080238401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267713 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9kfz\" (UniqueName: \"kubernetes.io/projected/94325ed6-8249-4f3a-98fc-fc507ab5ec18-kube-api-access-h9kfz\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267815 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-plugins-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267854 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267885 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-audit-policies\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267905 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8glqh\" (UniqueName: \"kubernetes.io/projected/1724971d-d75c-4e8c-a951-0f10e9c5e799-kube-api-access-8glqh\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267949 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/94325ed6-8249-4f3a-98fc-fc507ab5ec18-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.267981 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72ljv\" (UniqueName: \"kubernetes.io/projected/e8eefdff-74f7-4bb8-b39f-753a1375a403-kube-api-access-72ljv\") pod \"ingress-canary-46z5v\" (UID: \"e8eefdff-74f7-4bb8-b39f-753a1375a403\") " pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268007 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1724971d-d75c-4e8c-a951-0f10e9c5e799-audit-dir\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268031 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlwq6\" (UniqueName: \"kubernetes.io/projected/6cd2a062-18f8-44d1-918e-02142c6a2c16-kube-api-access-tlwq6\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268053 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/94325ed6-8249-4f3a-98fc-fc507ab5ec18-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268078 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncrfv\" (UniqueName: \"kubernetes.io/projected/26b0b618-116b-4aa2-8e84-3e024a2c46fd-kube-api-access-ncrfv\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268111 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-certs\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268135 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cd2a062-18f8-44d1-918e-02142c6a2c16-config-volume\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268351 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b0b618-116b-4aa2-8e84-3e024a2c46fd-config\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268374 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6rsb\" (UniqueName: \"kubernetes.io/projected/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-kube-api-access-w6rsb\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268405 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-registration-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268432 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268456 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d978f007-fe7a-4956-afc5-c83101961ee8-serving-cert\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268477 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-mountpoint-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268500 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26b0b618-116b-4aa2-8e84-3e024a2c46fd-auth-proxy-config\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268529 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-serving-cert\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268547 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6cd2a062-18f8-44d1-918e-02142c6a2c16-metrics-tls\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268571 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-service-ca-bundle\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268675 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26b0b618-116b-4aa2-8e84-3e024a2c46fd-machine-approver-tls\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268711 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4vl7\" (UniqueName: \"kubernetes.io/projected/d978f007-fe7a-4956-afc5-c83101961ee8-kube-api-access-n4vl7\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268733 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-node-bootstrap-token\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268754 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-socket-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268755 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268807 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-config\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.268972 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9gfq\" (UniqueName: \"kubernetes.io/projected/5921a039-7ee3-49ce-aad8-cf711e43e797-kube-api-access-f9gfq\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.269153 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-csi-data-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.269179 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/94325ed6-8249-4f3a-98fc-fc507ab5ec18-ready\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.269595 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-registration-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.269733 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-plugins-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.270011 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-mountpoint-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.270262 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-service-ca-bundle\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.271078 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/94325ed6-8249-4f3a-98fc-fc507ab5ec18-ready\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.271691 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-socket-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.271766 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.275552 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-audit-policies\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.276022 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/94325ed6-8249-4f3a-98fc-fc507ab5ec18-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.276097 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1724971d-d75c-4e8c-a951-0f10e9c5e799-audit-dir\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.276228 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26b0b618-116b-4aa2-8e84-3e024a2c46fd-machine-approver-tls\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.276835 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26b0b618-116b-4aa2-8e84-3e024a2c46fd-auth-proxy-config\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.277056 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-encryption-config\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.278826 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mvq8\" (UniqueName: \"kubernetes.io/projected/f1c2692c-98ff-4bd0-8c81-8249f124b464-kube-api-access-6mvq8\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.280374 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1724971d-d75c-4e8c-a951-0f10e9c5e799-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.281563 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d978f007-fe7a-4956-afc5-c83101961ee8-config\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.282077 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/94325ed6-8249-4f3a-98fc-fc507ab5ec18-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.282151 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5921a039-7ee3-49ce-aad8-cf711e43e797-csi-data-dir\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.282403 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cd2a062-18f8-44d1-918e-02142c6a2c16-config-volume\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.283164 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8eefdff-74f7-4bb8-b39f-753a1375a403-cert\") pod \"ingress-canary-46z5v\" (UID: \"e8eefdff-74f7-4bb8-b39f-753a1375a403\") " pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.284154 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-etcd-client\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.286145 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1724971d-d75c-4e8c-a951-0f10e9c5e799-serving-cert\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.290263 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d978f007-fe7a-4956-afc5-c83101961ee8-serving-cert\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.290962 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6cd2a062-18f8-44d1-918e-02142c6a2c16-metrics-tls\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.291447 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-certs\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.291713 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-node-bootstrap-token\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.296499 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg86s\" (UniqueName: \"kubernetes.io/projected/66202f9d-911e-47f7-b2d2-6326dd9c7283-kube-api-access-cg86s\") pod \"machine-config-operator-74547568cd-7xgrx\" (UID: \"66202f9d-911e-47f7-b2d2-6326dd9c7283\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.298409 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.343082 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.363423 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1c2692c-98ff-4bd0-8c81-8249f124b464-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jp4wl\" (UID: \"f1c2692c-98ff-4bd0-8c81-8249f124b464\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.366340 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7dd09639-861b-4f3d-b001-e28c8650cad7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4zfx7\" (UID: \"7dd09639-861b-4f3d-b001-e28c8650cad7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.366880 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.371053 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.371417 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:30.871400741 +0000 UTC m=+118.184011869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: W0314 09:12:30.384531 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod694288bd_c353_44f3_806b_5ce7c6a10486.slice/crio-17d725ca5f6039aa8030a40ad70cb91a6437f23193cbbce981e220c31e804174 WatchSource:0}: Error finding container 17d725ca5f6039aa8030a40ad70cb91a6437f23193cbbce981e220c31e804174: Status 404 returned error can't find the container with id 17d725ca5f6039aa8030a40ad70cb91a6437f23193cbbce981e220c31e804174 Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.385141 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e21032b1-6e5a-480c-92df-d60d83844fc2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.393787 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.406034 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mln7r\" (UniqueName: \"kubernetes.io/projected/e494d22d-e0ff-4fab-85d6-5241d72ff039-kube-api-access-mln7r\") pod \"olm-operator-6b444d44fb-29xcc\" (UID: \"e494d22d-e0ff-4fab-85d6-5241d72ff039\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.438747 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pnk2\" (UniqueName: \"kubernetes.io/projected/991014d6-3d2c-4291-91ef-db919b1c7d1c-kube-api-access-9pnk2\") pod \"marketplace-operator-79b997595-xtq2f\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.443441 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.444493 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wzqt\" (UniqueName: \"kubernetes.io/projected/ecf31fa9-e43c-4d64-95bd-4a52938b67f6-kube-api-access-7wzqt\") pod \"apiserver-76f77b778f-n64hl\" (UID: \"ecf31fa9-e43c-4d64-95bd-4a52938b67f6\") " pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.462251 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.465867 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.472875 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.473297 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:30.973282063 +0000 UTC m=+118.285893191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.473399 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.477482 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tgq2\" (UniqueName: \"kubernetes.io/projected/30f8e198-f784-4504-8fa3-09b59daa7b4d-kube-api-access-9tgq2\") pod \"openshift-controller-manager-operator-756b6f6bc6-5hr2h\" (UID: \"30f8e198-f784-4504-8fa3-09b59daa7b4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.491485 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.500519 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.501638 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.513991 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.517085 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jj2ff"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.517633 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj4wd\" (UniqueName: \"kubernetes.io/projected/e037c68c-36ae-47b2-b3f5-e159c512cf72-kube-api-access-cj4wd\") pod \"service-ca-operator-777779d784-dtnxl\" (UID: \"e037c68c-36ae-47b2-b3f5-e159c512cf72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.529317 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.529793 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g77xg\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-kube-api-access-g77xg\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.538384 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.552505 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:30 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:30 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:30 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.552556 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.554070 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swrkm\" (UniqueName: \"kubernetes.io/projected/35935053-e284-4215-b14c-efcb387201c4-kube-api-access-swrkm\") pod \"collect-profiles-29557980-gk2g6\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.555263 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnfxk\" (UniqueName: \"kubernetes.io/projected/c4b10c60-ed92-4bcf-9385-bf169e81051f-kube-api-access-nnfxk\") pod \"catalog-operator-68c6474976-z45r5\" (UID: \"c4b10c60-ed92-4bcf-9385-bf169e81051f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.571822 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgpzp\" (UniqueName: \"kubernetes.io/projected/e21032b1-6e5a-480c-92df-d60d83844fc2-kube-api-access-vgpzp\") pod \"cluster-image-registry-operator-dc59b4c8b-tnfbk\" (UID: \"e21032b1-6e5a-480c-92df-d60d83844fc2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.580717 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.580995 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.08098378 +0000 UTC m=+118.393594908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.583204 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.594349 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds9h2\" (UniqueName: \"kubernetes.io/projected/9fafc3ce-d8b1-4179-be51-d7d3f0c17217-kube-api-access-ds9h2\") pod \"multus-admission-controller-857f4d67dd-bfc6h\" (UID: \"9fafc3ce-d8b1-4179-be51-d7d3f0c17217\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.600905 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e938b24-b7e8-4f6e-9c4a-2a93410278a4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5v9l7\" (UID: \"6e938b24-b7e8-4f6e-9c4a-2a93410278a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.608091 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.619994 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15553d9e-d0a1-495b-8c4f-35f3d026fb33-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j5cd7\" (UID: \"15553d9e-d0a1-495b-8c4f-35f3d026fb33\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.622235 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.645627 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9kfz\" (UniqueName: \"kubernetes.io/projected/94325ed6-8249-4f3a-98fc-fc507ab5ec18-kube-api-access-h9kfz\") pod \"cni-sysctl-allowlist-ds-mrmxq\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.672161 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zldnt"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.675919 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4vl7\" (UniqueName: \"kubernetes.io/projected/d978f007-fe7a-4956-afc5-c83101961ee8-kube-api-access-n4vl7\") pod \"authentication-operator-69f744f599-nxkww\" (UID: \"d978f007-fe7a-4956-afc5-c83101961ee8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.681458 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.681844 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.18182385 +0000 UTC m=+118.494434988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.738310 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.740600 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8glqh\" (UniqueName: \"kubernetes.io/projected/1724971d-d75c-4e8c-a951-0f10e9c5e799-kube-api-access-8glqh\") pod \"apiserver-7bbb656c7d-4ftzb\" (UID: \"1724971d-d75c-4e8c-a951-0f10e9c5e799\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.743380 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72ljv\" (UniqueName: \"kubernetes.io/projected/e8eefdff-74f7-4bb8-b39f-753a1375a403-kube-api-access-72ljv\") pod \"ingress-canary-46z5v\" (UID: \"e8eefdff-74f7-4bb8-b39f-753a1375a403\") " pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.747431 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlwq6\" (UniqueName: \"kubernetes.io/projected/6cd2a062-18f8-44d1-918e-02142c6a2c16-kube-api-access-tlwq6\") pod \"dns-default-mr75n\" (UID: \"6cd2a062-18f8-44d1-918e-02142c6a2c16\") " pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.752164 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.754202 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncrfv\" (UniqueName: \"kubernetes.io/projected/26b0b618-116b-4aa2-8e84-3e024a2c46fd-kube-api-access-ncrfv\") pod \"machine-approver-56656f9798-fkgxr\" (UID: \"26b0b618-116b-4aa2-8e84-3e024a2c46fd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.756162 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9gfq\" (UniqueName: \"kubernetes.io/projected/5921a039-7ee3-49ce-aad8-cf711e43e797-kube-api-access-f9gfq\") pod \"csi-hostpathplugin-swrz2\" (UID: \"5921a039-7ee3-49ce-aad8-cf711e43e797\") " pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.759616 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.784119 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.784589 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.284551287 +0000 UTC m=+118.597162605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.788730 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6rsb\" (UniqueName: \"kubernetes.io/projected/d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5-kube-api-access-w6rsb\") pod \"machine-config-server-xv469\" (UID: \"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5\") " pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.835557 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.846079 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.875997 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.886628 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.891130 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.391089939 +0000 UTC m=+118.703701177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.949342 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.955308 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.961180 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.967322 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w"] Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.968409 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.985072 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-46z5v" Mar 14 09:12:30 crc kubenswrapper[4843]: I0314 09:12:30.987933 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:30 crc kubenswrapper[4843]: E0314 09:12:30.992143 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.492123345 +0000 UTC m=+118.804734473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.009953 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.020390 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xv469" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.040687 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.091368 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.091805 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.59178843 +0000 UTC m=+118.904399558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.119418 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" event={"ID":"1074b905-740e-4fc3-919b-e20309a9098b","Type":"ContainerStarted","Data":"04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.120157 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.125088 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" event={"ID":"694288bd-c353-44f3-806b-5ce7c6a10486","Type":"ContainerStarted","Data":"d81dafe2697bb50d002fbd17936b11e13c70975ce56376f4090b89d287909e37"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.125133 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" event={"ID":"694288bd-c353-44f3-806b-5ce7c6a10486","Type":"ContainerStarted","Data":"17d725ca5f6039aa8030a40ad70cb91a6437f23193cbbce981e220c31e804174"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.128169 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" event={"ID":"47b9849c-fc22-472c-a1f6-bd0e7a962a75","Type":"ContainerStarted","Data":"ba953c32584a27221dcf82b7636ac2082f6284abd9aaadf92f29793ac08d441a"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.128228 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" event={"ID":"47b9849c-fc22-472c-a1f6-bd0e7a962a75","Type":"ContainerStarted","Data":"d2c14c59ae56bf593f22d7fac8b5778c465320c7f6a4edf11acaccf9d417b57c"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.130452 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" event={"ID":"664b59b6-607d-4320-8802-0ce25a001245","Type":"ContainerStarted","Data":"39f35b2172e35d29da84d499355a75d44302b47bc74c4b8623de2d76dbeaff5d"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.145421 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" event={"ID":"dcf6a857-49d2-4319-b083-3ff05a19191b","Type":"ContainerStarted","Data":"7e750d1f6a96d30238cc860cd7c2b6bf148da7c014ce10c696c4794e06692a25"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.149572 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" event={"ID":"be6a1160-25cd-4a04-ad30-f1c40c8d7e05","Type":"ContainerStarted","Data":"4505d6b0f427c5b91319eebfc1177b889caf320108836b702578ac3a22a72a76"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.153663 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" event={"ID":"a46e0a56-02cc-43c0-b7ab-1c792b04fa2b","Type":"ContainerStarted","Data":"b260975aca65f370a66e444982555db5ca3a566cbaae1c2c6b9955ab5651c261"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.154645 4843 patch_prober.go:28] interesting pod/console-operator-58897d9998-ghbdz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.154681 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" podUID="a46e0a56-02cc-43c0-b7ab-1c792b04fa2b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.167386 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" event={"ID":"20cc2773-132c-4a74-9691-748a73111124","Type":"ContainerStarted","Data":"1f425c36cbd728ee4c161d1df531870b7ad634b845afa6414cab5937649a4a34"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.167729 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.171628 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2j8wl" event={"ID":"5c05075c-f721-4e78-ad5e-aaf88989336d","Type":"ContainerStarted","Data":"9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.176108 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" event={"ID":"4aba2d14-caf3-485c-8ee7-725ed508d1cf","Type":"ContainerStarted","Data":"37dd2ef1bead517dd2e49083f8d96bdb3e15f581ec60306e2680861dcab18607"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.184470 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" event={"ID":"bac1c5d6-a775-483f-ad66-20143a6b5141","Type":"ContainerStarted","Data":"3866973b24d16455e5b78997a668e196ffdae40c53c864addb50d7ffc80341d8"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.185182 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" event={"ID":"94325ed6-8249-4f3a-98fc-fc507ab5ec18","Type":"ContainerStarted","Data":"dd28d4d119a552603d8de09fbadb3b31ab48a3d59960c470a53c2524bfc299d6"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.189025 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" event={"ID":"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76","Type":"ContainerStarted","Data":"004df4f5a18185610086f5311f6d22d614ad2a74cbad30dc4cd578b0956aca1b"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.189058 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" event={"ID":"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76","Type":"ContainerStarted","Data":"59250cc1b9f5f3b7109e13f0678550a774bfaef6f6b2e8d2bed4d406e7d318f5"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.195734 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.196248 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.696235648 +0000 UTC m=+119.008846776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.204626 4843 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mt9q4 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.204701 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" podUID="1074b905-740e-4fc3-919b-e20309a9098b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.257630 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" event={"ID":"82f06b03-5a8e-4ef6-aac7-3c0b9326183e","Type":"ContainerStarted","Data":"58971c12d18c81187e49a05e048aa35c3f277777008d377918db715666bc9e91"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.257681 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" event={"ID":"82f06b03-5a8e-4ef6-aac7-3c0b9326183e","Type":"ContainerStarted","Data":"26506635b33dbf5f9a88d66eb6eb022841ad32dc3475650f2ef6c6e8c904d88d"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.266419 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" event={"ID":"8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6","Type":"ContainerStarted","Data":"46dfa55602f025ef5142b04aa819c4bba3b3a91f60705b4e1744848234151c8b"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.266501 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" event={"ID":"8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6","Type":"ContainerStarted","Data":"89a2157f5190df06d1549ea2360c936d9ccb7438838a123929979d0cb4e89697"} Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.266684 4843 patch_prober.go:28] interesting pod/downloads-7954f5f757-tr9zk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.266722 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tr9zk" podUID="570cf110-147b-482b-afe9-2b0ca133b6dc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.297008 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.298306 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.798260645 +0000 UTC m=+119.110871773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.358671 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" podStartSLOduration=45.358654305 podStartE2EDuration="45.358654305s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:31.356919433 +0000 UTC m=+118.669530561" watchObservedRunningTime="2026-03-14 09:12:31.358654305 +0000 UTC m=+118.671265433" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.359460 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.399227 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.399309 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.400524 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:31.900510247 +0000 UTC m=+119.213121375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.439094 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca968925-6d0f-46b7-b12b-7c617faabd3c-metrics-certs\") pod \"network-metrics-daemon-2zndz\" (UID: \"ca968925-6d0f-46b7-b12b-7c617faabd3c\") " pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.498595 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl"] Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.498625 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5kqvs"] Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.505762 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.506497 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.006474822 +0000 UTC m=+119.319085950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.506810 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.507605 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.007587366 +0000 UTC m=+119.320198494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.525476 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h"] Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.527704 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx"] Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.553406 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf"] Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.616047 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.616437 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.116422866 +0000 UTC m=+119.429033994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.643143 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:31 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:31 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:31 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.643436 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.679523 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2zndz" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.701801 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-pwxsw" podStartSLOduration=44.70177763 podStartE2EDuration="44.70177763s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:31.679300532 +0000 UTC m=+118.991911660" watchObservedRunningTime="2026-03-14 09:12:31.70177763 +0000 UTC m=+119.014388758" Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.718331 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.718636 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.218623278 +0000 UTC m=+119.531234406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.818870 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.819248 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.319228241 +0000 UTC m=+119.631839369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:31 crc kubenswrapper[4843]: I0314 09:12:31.921912 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:31 crc kubenswrapper[4843]: E0314 09:12:31.922539 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.422526475 +0000 UTC m=+119.735137603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.025611 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.026007 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.525993585 +0000 UTC m=+119.838604713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.085540 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-tr9zk" podStartSLOduration=46.085515269 podStartE2EDuration="46.085515269s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.078049734 +0000 UTC m=+119.390660862" watchObservedRunningTime="2026-03-14 09:12:32.085515269 +0000 UTC m=+119.398126397" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.087723 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n64hl"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.104954 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.126922 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.127255 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.627239257 +0000 UTC m=+119.939850385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: W0314 09:12:32.188468 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dd09639_861b_4f3d_b001_e28c8650cad7.slice/crio-ca21359dad55f6e3a0eb5f513ee14ed7f419b62c6cad6f27a624475d93c81839 WatchSource:0}: Error finding container ca21359dad55f6e3a0eb5f513ee14ed7f419b62c6cad6f27a624475d93c81839: Status 404 returned error can't find the container with id ca21359dad55f6e3a0eb5f513ee14ed7f419b62c6cad6f27a624475d93c81839 Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.200216 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2j8wl" podStartSLOduration=46.200190136 podStartE2EDuration="46.200190136s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.199257829 +0000 UTC m=+119.511868957" watchObservedRunningTime="2026-03-14 09:12:32.200190136 +0000 UTC m=+119.512801264" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.228968 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.229482 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.729460019 +0000 UTC m=+120.042071147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.315118 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" event={"ID":"30f8e198-f784-4504-8fa3-09b59daa7b4d","Type":"ContainerStarted","Data":"818ba839dc3c28675f78468c1f5e74c1d09d6dd5d82d849feb86d4d546409dc6"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.333032 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.333642 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.83362767 +0000 UTC m=+120.146238798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.360795 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" event={"ID":"664b59b6-607d-4320-8802-0ce25a001245","Type":"ContainerStarted","Data":"a25d079c2109343dc441f5552d4a894d4c543f2c9165b1c10ad955e4261aa8a3"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.384389 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.389819 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" event={"ID":"bac1c5d6-a775-483f-ad66-20143a6b5141","Type":"ContainerStarted","Data":"1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.389955 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.392344 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.396712 4843 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jj2ff container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.26:6443/healthz\": dial tcp 10.217.0.26:6443: connect: connection refused" start-of-body= Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.396781 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" podUID="664b59b6-607d-4320-8802-0ce25a001245" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.26:6443/healthz\": dial tcp 10.217.0.26:6443: connect: connection refused" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.400047 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.415010 4843 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4nzm6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.415211 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" podUID="bac1c5d6-a775-483f-ad66-20143a6b5141" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.415568 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" event={"ID":"26b0b618-116b-4aa2-8e84-3e024a2c46fd","Type":"ContainerStarted","Data":"b2f19a8f638640a72649b6f4e93b7525af108fcbed873b5e2962eded2c091756"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.423671 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" event={"ID":"8d64bc52-9c28-4efb-8f3f-aeeb8a738fc6","Type":"ContainerStarted","Data":"3cdfdae373fb681d6462d5afb69b6ceee8037625c727c8eff179f82615681af7"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.433998 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.434570 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.934549492 +0000 UTC m=+120.247160620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.434764 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.436614 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:32.936599344 +0000 UTC m=+120.249210472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.471313 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" event={"ID":"f1c2692c-98ff-4bd0-8c81-8249f124b464","Type":"ContainerStarted","Data":"68454778b73f84ccf81f36003382e2cb496b2ddc30375ca3c16e6eab9f749160"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.502324 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.510250 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.522778 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-f2fl7" podStartSLOduration=46.522595757 podStartE2EDuration="46.522595757s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.517247416 +0000 UTC m=+119.829858544" watchObservedRunningTime="2026-03-14 09:12:32.522595757 +0000 UTC m=+119.835206895" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.530622 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" event={"ID":"94325ed6-8249-4f3a-98fc-fc507ab5ec18","Type":"ContainerStarted","Data":"e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.531112 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.535394 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.536777 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" event={"ID":"8278bb86-cd39-4ecb-b0c6-75ec86c0763a","Type":"ContainerStarted","Data":"233734a149c85a18d48dc571dd7c2a428df93ede62b2f0fd4e8b43992aa6a511"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.536827 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" event={"ID":"8278bb86-cd39-4ecb-b0c6-75ec86c0763a","Type":"ContainerStarted","Data":"a25c1b75e0c2fd35a09fef8f93f5f2866f60adb0a52a33a202d9e56a248e7635"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.536977 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.537686 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.0376628 +0000 UTC m=+120.350274118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.549099 4843 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f5gt8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.549170 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:32 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:32 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:32 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.549202 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.549166 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" podUID="8278bb86-cd39-4ecb-b0c6-75ec86c0763a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.550111 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" event={"ID":"ecf31fa9-e43c-4d64-95bd-4a52938b67f6","Type":"ContainerStarted","Data":"30e42f4928c664037c79679e4d6f4ad130692f9f75c3e065959704bedeee64db"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.561669 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" event={"ID":"ce314766-eb86-4b0a-909b-2ae24850a4af","Type":"ContainerStarted","Data":"b2cb4300c31abda4e0549a6f73f259f07bda04af0d448a2ec0ebb1bf735ebf29"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.567853 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" event={"ID":"694288bd-c353-44f3-806b-5ce7c6a10486","Type":"ContainerStarted","Data":"40e2eb959fd6d60bf47a39684f872aa0e2acc33208dae12d9d9fd4bfaf5e32cb"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.568901 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.579636 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" event={"ID":"6e221699-471c-4cec-9aa8-b680923ac8b9","Type":"ContainerStarted","Data":"f5726a9b2f593a08f3d10d5498be7533809c10b5cb7c0e655c29a5ed87c0a3a3"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.597349 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" event={"ID":"66202f9d-911e-47f7-b2d2-6326dd9c7283","Type":"ContainerStarted","Data":"ca30f04440711f07346ecc00b948e9fecf50ddbf4acd949e1ff6046886857af4"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.603985 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.622117 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" event={"ID":"e6fc3d87-f84d-4782-8cab-adf072d642b4","Type":"ContainerStarted","Data":"b4868c939ee1896da2204935bcd78c59426fd021346b106195683c874dc14e3e"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.622184 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" event={"ID":"e6fc3d87-f84d-4782-8cab-adf072d642b4","Type":"ContainerStarted","Data":"8206625096c4fc8d75872ffeb30f6ec021b785ff2b95288b61ff09ffdd061b28"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.640066 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.641517 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.141500421 +0000 UTC m=+120.454111549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.648362 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-46z5v"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.648539 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qxscp" podStartSLOduration=45.648515702 podStartE2EDuration="45.648515702s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.636309275 +0000 UTC m=+119.948920403" watchObservedRunningTime="2026-03-14 09:12:32.648515702 +0000 UTC m=+119.961126830" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.654849 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" event={"ID":"fb768ff1-000e-4bef-9c37-fcfb4fdd7e76","Type":"ContainerStarted","Data":"b6151d354eb596f33972d54720e201090f7cd6f52841ecca394be82195bd8258"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.656905 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.660453 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nxkww"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.709055 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.714921 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" event={"ID":"7dd09639-861b-4f3d-b001-e28c8650cad7","Type":"ContainerStarted","Data":"ca21359dad55f6e3a0eb5f513ee14ed7f419b62c6cad6f27a624475d93c81839"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.717827 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" event={"ID":"dcf6a857-49d2-4319-b083-3ff05a19191b","Type":"ContainerStarted","Data":"7572cf713c92b348ed82cee73fdb3fcc3d4c1aa4cf82c4f2f8e3c6e689d30bbd"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.720608 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xv469" event={"ID":"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5","Type":"ContainerStarted","Data":"a9415d7e67ff2bc15a975e4ab1dfc119ea975b0bd7a3163aa7ff67bdf1d6b66b"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.720659 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xv469" event={"ID":"d11b2fa2-c2dc-4fe6-b195-d5da5e91f2b5","Type":"ContainerStarted","Data":"1a53010a4483f58e878f7abefa2a5d960130076ce7457473982b6ca95b72a9da"} Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.735151 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.736543 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.756618 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bfc6h"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.758195 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-ghbdz" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.761993 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.762230 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.262206541 +0000 UTC m=+120.574817669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.766249 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.770820 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.27080082 +0000 UTC m=+120.583411938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.786679 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mr75n"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.824711 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xtq2f"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.839147 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2zndz"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.839626 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cxr6c" podStartSLOduration=45.839615854 podStartE2EDuration="45.839615854s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.777217033 +0000 UTC m=+120.089828181" watchObservedRunningTime="2026-03-14 09:12:32.839615854 +0000 UTC m=+120.152226982" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.852496 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-swrz2"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.867291 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.867516 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.367492715 +0000 UTC m=+120.680103843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.867898 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.874627 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.374612859 +0000 UTC m=+120.687223987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.903060 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" podStartSLOduration=46.903040906 podStartE2EDuration="46.903040906s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.827786778 +0000 UTC m=+120.140397906" watchObservedRunningTime="2026-03-14 09:12:32.903040906 +0000 UTC m=+120.215652034" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.905338 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb"] Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.924560 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" podStartSLOduration=45.924533765 podStartE2EDuration="45.924533765s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.86103001 +0000 UTC m=+120.173641148" watchObservedRunningTime="2026-03-14 09:12:32.924533765 +0000 UTC m=+120.237144893" Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.926759 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vkqb9" podStartSLOduration=46.926749801 podStartE2EDuration="46.926749801s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.892570631 +0000 UTC m=+120.205181759" watchObservedRunningTime="2026-03-14 09:12:32.926749801 +0000 UTC m=+120.239360929" Mar 14 09:12:32 crc kubenswrapper[4843]: W0314 09:12:32.933152 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca968925_6d0f_46b7_b12b_7c617faabd3c.slice/crio-78577625c017e7135844fdbe06908e0bb7d869415e3bd8a7f83e8621fc40b9fb WatchSource:0}: Error finding container 78577625c017e7135844fdbe06908e0bb7d869415e3bd8a7f83e8621fc40b9fb: Status 404 returned error can't find the container with id 78577625c017e7135844fdbe06908e0bb7d869415e3bd8a7f83e8621fc40b9fb Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.973392 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:32 crc kubenswrapper[4843]: E0314 09:12:32.973773 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.473751188 +0000 UTC m=+120.786362316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:32 crc kubenswrapper[4843]: I0314 09:12:32.996659 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" podStartSLOduration=46.996640018 podStartE2EDuration="46.996640018s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.947291571 +0000 UTC m=+120.259902699" watchObservedRunningTime="2026-03-14 09:12:32.996640018 +0000 UTC m=+120.309251146" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.052691 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zkgfh" podStartSLOduration=46.052671957 podStartE2EDuration="46.052671957s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:32.997813604 +0000 UTC m=+120.310424732" watchObservedRunningTime="2026-03-14 09:12:33.052671957 +0000 UTC m=+120.365283085" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.053335 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" podStartSLOduration=47.053329138 podStartE2EDuration="47.053329138s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.050924025 +0000 UTC m=+120.363535153" watchObservedRunningTime="2026-03-14 09:12:33.053329138 +0000 UTC m=+120.365940266" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.080377 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.080677 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.580665311 +0000 UTC m=+120.893276439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.082055 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xv469" podStartSLOduration=6.082045673 podStartE2EDuration="6.082045673s" podCreationTimestamp="2026-03-14 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.081175806 +0000 UTC m=+120.393786934" watchObservedRunningTime="2026-03-14 09:12:33.082045673 +0000 UTC m=+120.394656801" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.128171 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" podStartSLOduration=46.128147713 podStartE2EDuration="46.128147713s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.1263845 +0000 UTC m=+120.438995628" watchObservedRunningTime="2026-03-14 09:12:33.128147713 +0000 UTC m=+120.440758841" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.159505 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" podStartSLOduration=46.159487767 podStartE2EDuration="46.159487767s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.158126996 +0000 UTC m=+120.470738134" watchObservedRunningTime="2026-03-14 09:12:33.159487767 +0000 UTC m=+120.472098895" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.180933 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.181169 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.681153721 +0000 UTC m=+120.993764849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.267747 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xvzgq" podStartSLOduration=46.267729011 podStartE2EDuration="46.267729011s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.234200501 +0000 UTC m=+120.546811629" watchObservedRunningTime="2026-03-14 09:12:33.267729011 +0000 UTC m=+120.580340139" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.296138 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" podStartSLOduration=46.296115807 podStartE2EDuration="46.296115807s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.26766388 +0000 UTC m=+120.580275008" watchObservedRunningTime="2026-03-14 09:12:33.296115807 +0000 UTC m=+120.608726935" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.296418 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.296752 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.796736015 +0000 UTC m=+121.109347143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.369970 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" podStartSLOduration=6.369950083 podStartE2EDuration="6.369950083s" podCreationTimestamp="2026-03-14 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.368754587 +0000 UTC m=+120.681365715" watchObservedRunningTime="2026-03-14 09:12:33.369950083 +0000 UTC m=+120.682561211" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.397292 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.397447 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.897416371 +0000 UTC m=+121.210027499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.397932 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.398236 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.898221935 +0000 UTC m=+121.210833063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.403968 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zldnt" podStartSLOduration=46.403949068 podStartE2EDuration="46.403949068s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.400664319 +0000 UTC m=+120.713275447" watchObservedRunningTime="2026-03-14 09:12:33.403949068 +0000 UTC m=+120.716560196" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.474451 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xjk7w" podStartSLOduration=47.474433883 podStartE2EDuration="47.474433883s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:33.426209129 +0000 UTC m=+120.738820257" watchObservedRunningTime="2026-03-14 09:12:33.474433883 +0000 UTC m=+120.787045011" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.498823 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.499122 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:33.999104337 +0000 UTC m=+121.311715465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.555678 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:33 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:33 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:33 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.555734 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.600656 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.600987 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.100975088 +0000 UTC m=+121.413586216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.701121 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.701476 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.201447046 +0000 UTC m=+121.514058174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.701920 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.722179 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.222161881 +0000 UTC m=+121.534773009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.776603 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" event={"ID":"991014d6-3d2c-4291-91ef-db919b1c7d1c","Type":"ContainerStarted","Data":"dc871c860e5aa687a79991225f75d3a6ab292336b24b55db833f6b9f7f9e3031"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.785667 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" event={"ID":"6e938b24-b7e8-4f6e-9c4a-2a93410278a4","Type":"ContainerStarted","Data":"b656d7d5a7e62d033726bce1fb2a6d1220f4be88e78ae3b24b62fea32e0b00fa"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.803834 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.804285 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.304246866 +0000 UTC m=+121.616857994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.808366 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" event={"ID":"66202f9d-911e-47f7-b2d2-6326dd9c7283","Type":"ContainerStarted","Data":"c1f69eff0629fc1a87f65ea9d52527acecaccbc6047854b529ffa23f1c4d3abf"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.808408 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" event={"ID":"66202f9d-911e-47f7-b2d2-6326dd9c7283","Type":"ContainerStarted","Data":"c03c7e310f82b6c6c2ea848df8fbf80c8d18f726f6f2e29a2d3542046bbe4e3d"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.812226 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" event={"ID":"7dd09639-861b-4f3d-b001-e28c8650cad7","Type":"ContainerStarted","Data":"e7a389a68d01d3ed8d5e0dc5fbe8f0e838e3cbfae00a6a468fb2dfe2dbf002c3"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.824561 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" event={"ID":"f1c2692c-98ff-4bd0-8c81-8249f124b464","Type":"ContainerStarted","Data":"cffd462f7171225806c6c12037a9fdc53b5e2da96f496de74f7a851801e1dffb"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.824616 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" event={"ID":"f1c2692c-98ff-4bd0-8c81-8249f124b464","Type":"ContainerStarted","Data":"6ef61afc790dea644eccc2930be48990e5d374bb159e7fc3cd0043ad8532d901"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.885657 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kd9zf" event={"ID":"6e221699-471c-4cec-9aa8-b680923ac8b9","Type":"ContainerStarted","Data":"f0c2962877664e57044c169bc70cfa5807031f96a3a5f31704c771ce193d809e"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.906973 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:33 crc kubenswrapper[4843]: E0314 09:12:33.909106 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.409090027 +0000 UTC m=+121.721701145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.940881 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" event={"ID":"26b0b618-116b-4aa2-8e84-3e024a2c46fd","Type":"ContainerStarted","Data":"919a0b8f28de9743a2af86e546ef6da1c9efa779e58b196bafb8193d85894321"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.940937 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" event={"ID":"26b0b618-116b-4aa2-8e84-3e024a2c46fd","Type":"ContainerStarted","Data":"635427e64fd5b32c26a2b637b9e1bb83d48f899bb5d918c7e9266df660a26a1a"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.944356 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mr75n" event={"ID":"6cd2a062-18f8-44d1-918e-02142c6a2c16","Type":"ContainerStarted","Data":"afe5cbfcf57f5f91ede08f60a614680a03559e6fd93e4b9f8bc89b455d0ec096"} Mar 14 09:12:33 crc kubenswrapper[4843]: I0314 09:12:33.993295 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" event={"ID":"5921a039-7ee3-49ce-aad8-cf711e43e797","Type":"ContainerStarted","Data":"ea86e65e790853bce10b28dc4eb69187eafeb7cc6fa68b2771a0069eed3525d5"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.013400 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.014655 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.514633509 +0000 UTC m=+121.827244637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.035215 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" event={"ID":"30f8e198-f784-4504-8fa3-09b59daa7b4d","Type":"ContainerStarted","Data":"00ce1411898ad40563aeee6d1cc9369013a29ff9ce28710f0dd6afcea61aceb7"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.100582 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" event={"ID":"ce314766-eb86-4b0a-909b-2ae24850a4af","Type":"ContainerStarted","Data":"7d5f40b7f7885912dd98f1a9ec87c961605777d6ec51fac96dec2a2c787076cc"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.100646 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" event={"ID":"ce314766-eb86-4b0a-909b-2ae24850a4af","Type":"ContainerStarted","Data":"2edc1b112ff7e76fbad69a70ea1658b8900d711f69ae97a9bbf9d7359879dbf0"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.118179 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" event={"ID":"9fafc3ce-d8b1-4179-be51-d7d3f0c17217","Type":"ContainerStarted","Data":"db3506dece26454714cfe3b8d4be1008d48130657cfdb779ee61a3cddf78afa4"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.119399 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.122464 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.62245211 +0000 UTC m=+121.935063238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.153439 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" event={"ID":"15553d9e-d0a1-495b-8c4f-35f3d026fb33","Type":"ContainerStarted","Data":"ddc860a5872e36fe8e6384d32b2e5a818ed3303d6c582ea1fb178b8d86b3bf6f"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.153489 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" event={"ID":"15553d9e-d0a1-495b-8c4f-35f3d026fb33","Type":"ContainerStarted","Data":"48b68a2aa47dfcdf023bba4674687305c32009a6c54fe8669865352d71837fe8"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.175149 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mrmxq"] Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.185569 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" event={"ID":"d978f007-fe7a-4956-afc5-c83101961ee8","Type":"ContainerStarted","Data":"63e6afaef1ef29c7a515686f8898d82fca6e2546965636d5ee95ae87dd6158e8"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.221161 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.222213 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" event={"ID":"e21032b1-6e5a-480c-92df-d60d83844fc2","Type":"ContainerStarted","Data":"8592c8c04cd85afbd84c10f7995c57c26bb49e4db43127e626ece9c3103b7738"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.222299 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" event={"ID":"e21032b1-6e5a-480c-92df-d60d83844fc2","Type":"ContainerStarted","Data":"679a1fbe0d05eb061095e1cb4195b4b0c63a020e5a839f06758d3da82f97f67e"} Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.222952 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.722934789 +0000 UTC m=+122.035545917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.259456 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" event={"ID":"35935053-e284-4215-b14c-efcb387201c4","Type":"ContainerStarted","Data":"cb175251cbd5d50cd19772b7f16f9388f5affe3b8c575fcc41ed78e061ce11a4"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.259498 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" event={"ID":"35935053-e284-4215-b14c-efcb387201c4","Type":"ContainerStarted","Data":"93ce632479c6fab2da361b1d93f19439fe8711ec899b0468b10436efd8fad0b2"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.293856 4843 generic.go:334] "Generic (PLEG): container finished" podID="ecf31fa9-e43c-4d64-95bd-4a52938b67f6" containerID="859ad766e7d49cd97c340e63464aa2cc3656be4257deb79ad4c91de5ed477088" exitCode=0 Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.294307 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" event={"ID":"ecf31fa9-e43c-4d64-95bd-4a52938b67f6","Type":"ContainerDied","Data":"859ad766e7d49cd97c340e63464aa2cc3656be4257deb79ad4c91de5ed477088"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.303559 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" event={"ID":"1724971d-d75c-4e8c-a951-0f10e9c5e799","Type":"ContainerStarted","Data":"deb3073bf404c564708b49cc9835dd754903e7869106007434bbac68b4416ca6"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.323750 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.326232 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.826215083 +0000 UTC m=+122.138826211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.330288 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" event={"ID":"e494d22d-e0ff-4fab-85d6-5241d72ff039","Type":"ContainerStarted","Data":"59d0b317693fb693cb9181f47dd59dba9ef36d27302fad797ed452a1f0ed761c"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.330344 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" event={"ID":"e494d22d-e0ff-4fab-85d6-5241d72ff039","Type":"ContainerStarted","Data":"ee16ba1c2c167a84068a3db716bd5fe4af214ade5fdf03e0aea2aedf9f2c356e"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.331045 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.342528 4843 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-29xcc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.342585 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" podUID="e494d22d-e0ff-4fab-85d6-5241d72ff039" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.377731 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-46z5v" event={"ID":"e8eefdff-74f7-4bb8-b39f-753a1375a403","Type":"ContainerStarted","Data":"9c5c6946889950f255d5c73295e762a10fcd16feaa212727038b05136b35d18c"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.388727 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.416084 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f29b51f4b235799e889cec4242cf9895f2ac9b154a1e4e1a4eaaea4ae9d4366e"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.416866 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.429402 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" event={"ID":"e037c68c-36ae-47b2-b3f5-e159c512cf72","Type":"ContainerStarted","Data":"df03aeda06eda15465621bc92875ccbf399712ee57386bcbb53d30e2c783018f"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.430327 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.431163 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:34.931140236 +0000 UTC m=+122.243751364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.448050 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" podStartSLOduration=47.448029936 podStartE2EDuration="47.448029936s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.447942293 +0000 UTC m=+121.760553441" watchObservedRunningTime="2026-03-14 09:12:34.448029936 +0000 UTC m=+121.760641064" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.481552 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" event={"ID":"c4b10c60-ed92-4bcf-9385-bf169e81051f","Type":"ContainerStarted","Data":"e4a0944bb44af86fc204ebbc84fb50117e1ffa51298247bc921bf0da5a6a3d5e"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.484180 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.502046 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2zndz" event={"ID":"ca968925-6d0f-46b7-b12b-7c617faabd3c","Type":"ContainerStarted","Data":"78577625c017e7135844fdbe06908e0bb7d869415e3bd8a7f83e8621fc40b9fb"} Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.542472 4843 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z45r5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.542540 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" podUID="c4b10c60-ed92-4bcf-9385-bf169e81051f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.543257 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.546829 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7xgrx" podStartSLOduration=47.546815473 podStartE2EDuration="47.546815473s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.503974732 +0000 UTC m=+121.816585860" watchObservedRunningTime="2026-03-14 09:12:34.546815473 +0000 UTC m=+121.859426601" Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.547765 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.047745902 +0000 UTC m=+122.360357030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.549157 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.554821 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:34 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:34 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:34 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.554916 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.570754 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ph27v" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.581686 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.605186 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j5cd7" podStartSLOduration=47.605160863 podStartE2EDuration="47.605160863s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.593649926 +0000 UTC m=+121.906261054" watchObservedRunningTime="2026-03-14 09:12:34.605160863 +0000 UTC m=+121.917771981" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.649241 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgxr" podStartSLOduration=48.649225791 podStartE2EDuration="48.649225791s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.646939482 +0000 UTC m=+121.959550600" watchObservedRunningTime="2026-03-14 09:12:34.649225791 +0000 UTC m=+121.961836919" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.665792 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.666992 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.166976216 +0000 UTC m=+122.479587344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.750337 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-46z5v" podStartSLOduration=7.750317999 podStartE2EDuration="7.750317999s" podCreationTimestamp="2026-03-14 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.690949019 +0000 UTC m=+122.003560147" watchObservedRunningTime="2026-03-14 09:12:34.750317999 +0000 UTC m=+122.062929127" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.750461 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tnfbk" podStartSLOduration=48.750457993 podStartE2EDuration="48.750457993s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.7486964 +0000 UTC m=+122.061307528" watchObservedRunningTime="2026-03-14 09:12:34.750457993 +0000 UTC m=+122.063069121" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.779100 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.779611 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.279590132 +0000 UTC m=+122.592201260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.788045 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5kqvs" podStartSLOduration=47.788023376 podStartE2EDuration="47.788023376s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.786654875 +0000 UTC m=+122.099266013" watchObservedRunningTime="2026-03-14 09:12:34.788023376 +0000 UTC m=+122.100634504" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.848365 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4zfx7" podStartSLOduration=47.848345345 podStartE2EDuration="47.848345345s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.842749136 +0000 UTC m=+122.155360264" watchObservedRunningTime="2026-03-14 09:12:34.848345345 +0000 UTC m=+122.160956473" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.886081 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.886479 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.386459984 +0000 UTC m=+122.699071112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.889763 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" podStartSLOduration=48.889739072 podStartE2EDuration="48.889739072s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.881494733 +0000 UTC m=+122.194105861" watchObservedRunningTime="2026-03-14 09:12:34.889739072 +0000 UTC m=+122.202350200" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.967283 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jp4wl" podStartSLOduration=47.967242979 podStartE2EDuration="47.967242979s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.931533892 +0000 UTC m=+122.244145020" watchObservedRunningTime="2026-03-14 09:12:34.967242979 +0000 UTC m=+122.279854107" Mar 14 09:12:34 crc kubenswrapper[4843]: I0314 09:12:34.990909 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:34 crc kubenswrapper[4843]: E0314 09:12:34.991244 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.491233492 +0000 UTC m=+122.803844620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.000689 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" podStartSLOduration=49.000671936 podStartE2EDuration="49.000671936s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:34.967682912 +0000 UTC m=+122.280294040" watchObservedRunningTime="2026-03-14 09:12:35.000671936 +0000 UTC m=+122.313283064" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.057945 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5hr2h" podStartSLOduration=49.057925542 podStartE2EDuration="49.057925542s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.000553103 +0000 UTC m=+122.313164241" watchObservedRunningTime="2026-03-14 09:12:35.057925542 +0000 UTC m=+122.370536670" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.093030 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.093400 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.593380682 +0000 UTC m=+122.905991810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.154893 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.154872656 podStartE2EDuration="18.154872656s" podCreationTimestamp="2026-03-14 09:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.120688045 +0000 UTC m=+122.433299183" watchObservedRunningTime="2026-03-14 09:12:35.154872656 +0000 UTC m=+122.467483784" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.189484 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" podStartSLOduration=48.189462748 podStartE2EDuration="48.189462748s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.185588181 +0000 UTC m=+122.498199309" watchObservedRunningTime="2026-03-14 09:12:35.189462748 +0000 UTC m=+122.502073876" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.194651 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.194657 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" podStartSLOduration=48.194629184 podStartE2EDuration="48.194629184s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.154756842 +0000 UTC m=+122.467367960" watchObservedRunningTime="2026-03-14 09:12:35.194629184 +0000 UTC m=+122.507240312" Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.195031 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.695017716 +0000 UTC m=+123.007628844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.255309 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f5gt8" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.297830 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.298042 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.798014881 +0000 UTC m=+123.110626009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.298471 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.298890 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.798878077 +0000 UTC m=+123.111489205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.402785 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.403429 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.903411769 +0000 UTC m=+123.216022897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.403490 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.403781 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:35.903774419 +0000 UTC m=+123.216385547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.513223 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.514957 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.014938001 +0000 UTC m=+123.327549129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.518812 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mr75n" event={"ID":"6cd2a062-18f8-44d1-918e-02142c6a2c16","Type":"ContainerStarted","Data":"9605e7688f081a3d115975dd7878caa26c0d21a5a2539c0ca530ae38f174159e"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.518879 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mr75n" event={"ID":"6cd2a062-18f8-44d1-918e-02142c6a2c16","Type":"ContainerStarted","Data":"82d57b0ded83920c3be3d0116bcc024c65d964a280287350ce456f591c5214e2"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.520174 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.524036 4843 generic.go:334] "Generic (PLEG): container finished" podID="35935053-e284-4215-b14c-efcb387201c4" containerID="cb175251cbd5d50cd19772b7f16f9388f5affe3b8c575fcc41ed78e061ce11a4" exitCode=0 Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.524127 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" event={"ID":"35935053-e284-4215-b14c-efcb387201c4","Type":"ContainerDied","Data":"cb175251cbd5d50cd19772b7f16f9388f5affe3b8c575fcc41ed78e061ce11a4"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.547078 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-46z5v" event={"ID":"e8eefdff-74f7-4bb8-b39f-753a1375a403","Type":"ContainerStarted","Data":"3edacb4f79e2829beb5cfc342a97fca1cc7818fd375bf1bb433c816cb23d7c10"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.565368 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:35 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:35 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:35 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.565413 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.584706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" event={"ID":"991014d6-3d2c-4291-91ef-db919b1c7d1c","Type":"ContainerStarted","Data":"98cb6d370b35bba7252e559ee327df9e2cac39087e0bfac591d32a12ba827dd6"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.585882 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.602410 4843 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xtq2f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/healthz\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.602636 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.13:8080/healthz\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.616886 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.617233 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.117220955 +0000 UTC m=+123.429832083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.621920 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" event={"ID":"6e938b24-b7e8-4f6e-9c4a-2a93410278a4","Type":"ContainerStarted","Data":"4c9942f2a9434d81ae336a4a8d9e99749b3d818279d86c77d987d9e54d006fe8"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.645425 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" event={"ID":"5921a039-7ee3-49ce-aad8-cf711e43e797","Type":"ContainerStarted","Data":"1505bd9948054eab3132bcece17afd5ddf5a66833a7c5c5a9db49b46e551b742"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.647573 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-mr75n" podStartSLOduration=8.647560669 podStartE2EDuration="8.647560669s" podCreationTimestamp="2026-03-14 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.602183801 +0000 UTC m=+122.914794929" watchObservedRunningTime="2026-03-14 09:12:35.647560669 +0000 UTC m=+122.960171797" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.648108 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" podStartSLOduration=48.648101876 podStartE2EDuration="48.648101876s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.645130796 +0000 UTC m=+122.957741924" watchObservedRunningTime="2026-03-14 09:12:35.648101876 +0000 UTC m=+122.960713004" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.666520 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2zndz" event={"ID":"ca968925-6d0f-46b7-b12b-7c617faabd3c","Type":"ContainerStarted","Data":"b23dbe0d731961d6172a23883ada4b106d89bfb6991d03a131ce4d6bcffd8b1b"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.666570 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2zndz" event={"ID":"ca968925-6d0f-46b7-b12b-7c617faabd3c","Type":"ContainerStarted","Data":"68632ce458cbec49e88a42ed9c861b18b31743e7522c05dbe306ccca7afab2c9"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.667452 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35238: no serving certificate available for the kubelet" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.684653 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" event={"ID":"c4b10c60-ed92-4bcf-9385-bf169e81051f","Type":"ContainerStarted","Data":"73c691d3092c7e46a5641eff79a499156d450d482367bd94862a8ed9d4e95fe7"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.692803 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" event={"ID":"ecf31fa9-e43c-4d64-95bd-4a52938b67f6","Type":"ContainerStarted","Data":"04d381c0ecf684cbb99b5e0701ccb823927d526a23d0bedd38771a53d16a5aba"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.703868 4843 generic.go:334] "Generic (PLEG): container finished" podID="1724971d-d75c-4e8c-a951-0f10e9c5e799" containerID="7d60b86bb1fafe89f0c61ad11e80568c33adfdc401ecb4dd3065c7de8d90b3c7" exitCode=0 Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.703952 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" event={"ID":"1724971d-d75c-4e8c-a951-0f10e9c5e799","Type":"ContainerDied","Data":"7d60b86bb1fafe89f0c61ad11e80568c33adfdc401ecb4dd3065c7de8d90b3c7"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.717654 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.718081 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nxkww" event={"ID":"d978f007-fe7a-4956-afc5-c83101961ee8","Type":"ContainerStarted","Data":"8caf10b2935f9af092dfe4bbb07ee85bb43aaa1833a256852a02a84cb40dd2db"} Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.719887 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.219869379 +0000 UTC m=+123.532480507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.725131 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z45r5" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.732525 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" event={"ID":"9fafc3ce-d8b1-4179-be51-d7d3f0c17217","Type":"ContainerStarted","Data":"6abce1d3a6efee13ee3b4c4ac5a707990bf7348749ecebd701554d5c00374045"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.732567 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" event={"ID":"9fafc3ce-d8b1-4179-be51-d7d3f0c17217","Type":"ContainerStarted","Data":"de1d239e3c8f5928d54ffbc6584a01f89009ca88496e1c1c5a26ef87655a431d"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.741070 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dtnxl" event={"ID":"e037c68c-36ae-47b2-b3f5-e159c512cf72","Type":"ContainerStarted","Data":"900afa3556f8617b24211d16e7c6615039a17b49581bdc24f79576caa58d5c83"} Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.741593 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" gracePeriod=30 Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.746155 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35250: no serving certificate available for the kubelet" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.750169 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2zndz" podStartSLOduration=49.750152593 podStartE2EDuration="49.750152593s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.724764467 +0000 UTC m=+123.037375595" watchObservedRunningTime="2026-03-14 09:12:35.750152593 +0000 UTC m=+123.062763711" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.751125 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5v9l7" podStartSLOduration=48.751120101 podStartE2EDuration="48.751120101s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.74907685 +0000 UTC m=+123.061687978" watchObservedRunningTime="2026-03-14 09:12:35.751120101 +0000 UTC m=+123.063731229" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.788897 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29xcc" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.820235 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.840748 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.340731603 +0000 UTC m=+123.653342731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.857662 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35256: no serving certificate available for the kubelet" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.922311 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:35 crc kubenswrapper[4843]: E0314 09:12:35.923131 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.423114477 +0000 UTC m=+123.735725605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.941497 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-bfc6h" podStartSLOduration=48.94147856 podStartE2EDuration="48.94147856s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:35.86982094 +0000 UTC m=+123.182432088" watchObservedRunningTime="2026-03-14 09:12:35.94147856 +0000 UTC m=+123.254089688" Mar 14 09:12:35 crc kubenswrapper[4843]: I0314 09:12:35.971496 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35262: no serving certificate available for the kubelet" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.024051 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.024376 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.52436419 +0000 UTC m=+123.836975308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.072867 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35278: no serving certificate available for the kubelet" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.124651 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.124967 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.624941231 +0000 UTC m=+123.937552359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.140053 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mt9q4"] Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.147925 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6"] Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.213596 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35294: no serving certificate available for the kubelet" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.233658 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.233985 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.733972099 +0000 UTC m=+124.046583227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.289474 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35310: no serving certificate available for the kubelet" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.334421 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.334577 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.834543901 +0000 UTC m=+124.147155029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.334666 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.335020 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.835009055 +0000 UTC m=+124.147620193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.431847 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35326: no serving certificate available for the kubelet" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.436418 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.436627 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.936596818 +0000 UTC m=+124.249207936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.436824 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.437170 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:36.937156774 +0000 UTC m=+124.249767902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.540184 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.540673 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.040655385 +0000 UTC m=+124.353266513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.540828 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:36 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:36 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:36 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.540901 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.540947 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.541249 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.041242263 +0000 UTC m=+124.353853391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.642152 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.642366 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.142338351 +0000 UTC m=+124.454949479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.642672 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.642993 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.142979189 +0000 UTC m=+124.455590317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.743327 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.743513 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.2434834 +0000 UTC m=+124.556094528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.743669 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.743973 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.243960054 +0000 UTC m=+124.556571182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.752008 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" event={"ID":"ecf31fa9-e43c-4d64-95bd-4a52938b67f6","Type":"ContainerStarted","Data":"fc6ddef2636069dee50f96d017af17d89e53fcbfb42ff44c4ee6252823c25aab"} Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.754892 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" event={"ID":"1724971d-d75c-4e8c-a951-0f10e9c5e799","Type":"ContainerStarted","Data":"64bed0947ddc66a3e58af99912486f398db68f41495b8fcd1094b020a2329275"} Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.757254 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" event={"ID":"5921a039-7ee3-49ce-aad8-cf711e43e797","Type":"ContainerStarted","Data":"8cb03b6fa36958b1c01c50892f845cbf941cd05793576db507bce9f3b074a84d"} Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.757297 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" event={"ID":"5921a039-7ee3-49ce-aad8-cf711e43e797","Type":"ContainerStarted","Data":"a160ddf2df40f0c92b3e21d703c5779d0fa654c1a26359b0f746b9c716ad0037"} Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.757389 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" podUID="1074b905-740e-4fc3-919b-e20309a9098b" containerName="controller-manager" containerID="cri-o://04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc" gracePeriod=30 Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.814857 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.845376 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.846583 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.346556318 +0000 UTC m=+124.659167446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.856574 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" podStartSLOduration=49.856554519 podStartE2EDuration="49.856554519s" podCreationTimestamp="2026-03-14 09:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:36.8556163 +0000 UTC m=+124.168227428" watchObservedRunningTime="2026-03-14 09:12:36.856554519 +0000 UTC m=+124.169165647" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.857092 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" podStartSLOduration=50.857087375 podStartE2EDuration="50.857087375s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:36.800683474 +0000 UTC m=+124.113294602" watchObservedRunningTime="2026-03-14 09:12:36.857087375 +0000 UTC m=+124.169698503" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.913469 4843 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 14 09:12:36 crc kubenswrapper[4843]: I0314 09:12:36.947014 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:36 crc kubenswrapper[4843]: E0314 09:12:36.947358 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.447346036 +0000 UTC m=+124.759957164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.047940 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.547923809 +0000 UTC m=+124.860534937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.047960 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.048214 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.048494 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.548485305 +0000 UTC m=+124.861096433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.089401 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35342: no serving certificate available for the kubelet" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.134123 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.149989 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.150469 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.65045083 +0000 UTC m=+124.963061968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.246356 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.251431 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35935053-e284-4215-b14c-efcb387201c4-config-volume\") pod \"35935053-e284-4215-b14c-efcb387201c4\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.251492 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swrkm\" (UniqueName: \"kubernetes.io/projected/35935053-e284-4215-b14c-efcb387201c4-kube-api-access-swrkm\") pod \"35935053-e284-4215-b14c-efcb387201c4\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.251760 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35935053-e284-4215-b14c-efcb387201c4-secret-volume\") pod \"35935053-e284-4215-b14c-efcb387201c4\" (UID: \"35935053-e284-4215-b14c-efcb387201c4\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.251870 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.251988 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35935053-e284-4215-b14c-efcb387201c4-config-volume" (OuterVolumeSpecName: "config-volume") pod "35935053-e284-4215-b14c-efcb387201c4" (UID: "35935053-e284-4215-b14c-efcb387201c4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.252227 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.752213747 +0000 UTC m=+125.064824875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.259117 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35935053-e284-4215-b14c-efcb387201c4-kube-api-access-swrkm" (OuterVolumeSpecName: "kube-api-access-swrkm") pod "35935053-e284-4215-b14c-efcb387201c4" (UID: "35935053-e284-4215-b14c-efcb387201c4"). InnerVolumeSpecName "kube-api-access-swrkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.259165 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35935053-e284-4215-b14c-efcb387201c4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "35935053-e284-4215-b14c-efcb387201c4" (UID: "35935053-e284-4215-b14c-efcb387201c4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.359448 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mss9d"] Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.359692 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1074b905-740e-4fc3-919b-e20309a9098b" containerName="controller-manager" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.359707 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1074b905-740e-4fc3-919b-e20309a9098b" containerName="controller-manager" Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.359723 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35935053-e284-4215-b14c-efcb387201c4" containerName="collect-profiles" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.359730 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="35935053-e284-4215-b14c-efcb387201c4" containerName="collect-profiles" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.359824 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1074b905-740e-4fc3-919b-e20309a9098b" containerName="controller-manager" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.359840 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="35935053-e284-4215-b14c-efcb387201c4" containerName="collect-profiles" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.360569 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.360793 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.360894 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.860870173 +0000 UTC m=+125.173481311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.361011 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9469\" (UniqueName: \"kubernetes.io/projected/1074b905-740e-4fc3-919b-e20309a9098b-kube-api-access-p9469\") pod \"1074b905-740e-4fc3-919b-e20309a9098b\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.361135 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-proxy-ca-bundles\") pod \"1074b905-740e-4fc3-919b-e20309a9098b\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.361289 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1074b905-740e-4fc3-919b-e20309a9098b-serving-cert\") pod \"1074b905-740e-4fc3-919b-e20309a9098b\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.361407 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-client-ca\") pod \"1074b905-740e-4fc3-919b-e20309a9098b\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.361508 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-config\") pod \"1074b905-740e-4fc3-919b-e20309a9098b\" (UID: \"1074b905-740e-4fc3-919b-e20309a9098b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.362753 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-config" (OuterVolumeSpecName: "config") pod "1074b905-740e-4fc3-919b-e20309a9098b" (UID: "1074b905-740e-4fc3-919b-e20309a9098b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.363299 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.363591 4843 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35935053-e284-4215-b14c-efcb387201c4-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.363620 4843 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35935053-e284-4215-b14c-efcb387201c4-config-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.363631 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.363640 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swrkm\" (UniqueName: \"kubernetes.io/projected/35935053-e284-4215-b14c-efcb387201c4-kube-api-access-swrkm\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.363946 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.863928415 +0000 UTC m=+125.176539543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.364857 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1074b905-740e-4fc3-919b-e20309a9098b" (UID: "1074b905-740e-4fc3-919b-e20309a9098b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.367482 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mss9d"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.368621 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1074b905-740e-4fc3-919b-e20309a9098b-kube-api-access-p9469" (OuterVolumeSpecName: "kube-api-access-p9469") pod "1074b905-740e-4fc3-919b-e20309a9098b" (UID: "1074b905-740e-4fc3-919b-e20309a9098b"). InnerVolumeSpecName "kube-api-access-p9469". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.369713 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-client-ca" (OuterVolumeSpecName: "client-ca") pod "1074b905-740e-4fc3-919b-e20309a9098b" (UID: "1074b905-740e-4fc3-919b-e20309a9098b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.370119 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1074b905-740e-4fc3-919b-e20309a9098b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1074b905-740e-4fc3-919b-e20309a9098b" (UID: "1074b905-740e-4fc3-919b-e20309a9098b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.370805 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.464655 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.464796 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.964771815 +0000 UTC m=+125.277382943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465221 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465365 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-catalog-content\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465518 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g98sp\" (UniqueName: \"kubernetes.io/projected/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-kube-api-access-g98sp\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465666 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-utilities\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465734 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9469\" (UniqueName: \"kubernetes.io/projected/1074b905-740e-4fc3-919b-e20309a9098b-kube-api-access-p9469\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465746 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465754 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1074b905-740e-4fc3-919b-e20309a9098b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.465763 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1074b905-740e-4fc3-919b-e20309a9098b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.465872 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-14 09:12:37.965864709 +0000 UTC m=+125.278475837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hfz6k" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.500521 4843 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-14T09:12:36.913873737Z","Handler":null,"Name":""} Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.504332 4843 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.504382 4843 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.542892 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:37 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:37 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:37 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.542971 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.546376 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kzj4r"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.547455 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.552912 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.554500 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzj4r"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.566909 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.567150 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-utilities\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.567240 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-catalog-content\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.567311 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g98sp\" (UniqueName: \"kubernetes.io/projected/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-kube-api-access-g98sp\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.567735 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-utilities\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.567765 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-catalog-content\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.584180 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.593330 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g98sp\" (UniqueName: \"kubernetes.io/projected/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-kube-api-access-g98sp\") pod \"certified-operators-mss9d\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.668950 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.669032 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-catalog-content\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.669086 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-utilities\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.669137 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhsjs\" (UniqueName: \"kubernetes.io/projected/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-kube-api-access-xhsjs\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.673734 4843 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.673783 4843 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.686928 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.713676 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hfz6k\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.745217 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zx9cb"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.746478 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.754402 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zx9cb"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.767596 4843 generic.go:334] "Generic (PLEG): container finished" podID="1074b905-740e-4fc3-919b-e20309a9098b" containerID="04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc" exitCode=0 Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.767731 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" event={"ID":"1074b905-740e-4fc3-919b-e20309a9098b","Type":"ContainerDied","Data":"04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc"} Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.767743 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.767769 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mt9q4" event={"ID":"1074b905-740e-4fc3-919b-e20309a9098b","Type":"ContainerDied","Data":"5f7c4a1b7081a880f02daf9170c25ef705ca33472782d639502359e9c32e769e"} Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.767809 4843 scope.go:117] "RemoveContainer" containerID="04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.767924 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.769873 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhsjs\" (UniqueName: \"kubernetes.io/projected/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-kube-api-access-xhsjs\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.769970 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-catalog-content\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.770011 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-utilities\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.771038 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-utilities\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.771643 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-catalog-content\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.777680 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" event={"ID":"35935053-e284-4215-b14c-efcb387201c4","Type":"ContainerDied","Data":"93ce632479c6fab2da361b1d93f19439fe8711ec899b0468b10436efd8fad0b2"} Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.777747 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93ce632479c6fab2da361b1d93f19439fe8711ec899b0468b10436efd8fad0b2" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.777883 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.778177 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.793999 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" event={"ID":"5921a039-7ee3-49ce-aad8-cf711e43e797","Type":"ContainerStarted","Data":"97bf8ff84d4a27ed66a72490f0142afdea3a0d77792fa8b1650a718ac6826a38"} Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.795198 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" podUID="bac1c5d6-a775-483f-ad66-20143a6b5141" containerName="route-controller-manager" containerID="cri-o://1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e" gracePeriod=30 Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.818926 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhsjs\" (UniqueName: \"kubernetes.io/projected/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-kube-api-access-xhsjs\") pod \"community-operators-kzj4r\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.833099 4843 scope.go:117] "RemoveContainer" containerID="04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc" Mar 14 09:12:37 crc kubenswrapper[4843]: E0314 09:12:37.833689 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc\": container with ID starting with 04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc not found: ID does not exist" containerID="04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.833721 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc"} err="failed to get container status \"04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc\": rpc error: code = NotFound desc = could not find container \"04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc\": container with ID starting with 04ceab48abb515e0c18c5bcbcd97c9cab15a02d25b1a68387cc65e019e3fb2dc not found: ID does not exist" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.836923 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-swrz2" podStartSLOduration=10.836903405 podStartE2EDuration="10.836903405s" podCreationTimestamp="2026-03-14 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:37.836434731 +0000 UTC m=+125.149045879" watchObservedRunningTime="2026-03-14 09:12:37.836903405 +0000 UTC m=+125.149514543" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.853660 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mt9q4"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.858505 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mt9q4"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.870029 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.871006 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-utilities\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.871156 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-catalog-content\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.871195 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lhj4\" (UniqueName: \"kubernetes.io/projected/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-kube-api-access-7lhj4\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.945567 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vrzt2"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.946611 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.988500 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrzt2"] Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.988889 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-utilities\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.989122 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-catalog-content\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.989166 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lhj4\" (UniqueName: \"kubernetes.io/projected/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-kube-api-access-7lhj4\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.989520 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-utilities\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.990555 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-catalog-content\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:37 crc kubenswrapper[4843]: I0314 09:12:37.995483 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mss9d"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.027425 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lhj4\" (UniqueName: \"kubernetes.io/projected/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-kube-api-access-7lhj4\") pod \"certified-operators-zx9cb\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.068325 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.093234 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-utilities\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.093328 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-catalog-content\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.093411 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krq99\" (UniqueName: \"kubernetes.io/projected/f36123e7-f666-4382-8324-252e94a27884-kube-api-access-krq99\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.147782 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hfz6k"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.194325 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-utilities\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.194381 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-catalog-content\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.194455 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krq99\" (UniqueName: \"kubernetes.io/projected/f36123e7-f666-4382-8324-252e94a27884-kube-api-access-krq99\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.195336 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-utilities\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.195636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-catalog-content\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.217649 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krq99\" (UniqueName: \"kubernetes.io/projected/f36123e7-f666-4382-8324-252e94a27884-kube-api-access-krq99\") pod \"community-operators-vrzt2\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.296932 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.311094 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zx9cb"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.346929 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.399702 4843 ???:1] "http: TLS handshake error from 192.168.126.11:35354: no serving certificate available for the kubelet" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.435939 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzj4r"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.500056 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca\") pod \"bac1c5d6-a775-483f-ad66-20143a6b5141\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.500156 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config\") pod \"bac1c5d6-a775-483f-ad66-20143a6b5141\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.500207 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbctp\" (UniqueName: \"kubernetes.io/projected/bac1c5d6-a775-483f-ad66-20143a6b5141-kube-api-access-zbctp\") pod \"bac1c5d6-a775-483f-ad66-20143a6b5141\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.500350 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert\") pod \"bac1c5d6-a775-483f-ad66-20143a6b5141\" (UID: \"bac1c5d6-a775-483f-ad66-20143a6b5141\") " Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.500910 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca" (OuterVolumeSpecName: "client-ca") pod "bac1c5d6-a775-483f-ad66-20143a6b5141" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.501977 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config" (OuterVolumeSpecName: "config") pod "bac1c5d6-a775-483f-ad66-20143a6b5141" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.507655 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac1c5d6-a775-483f-ad66-20143a6b5141-kube-api-access-zbctp" (OuterVolumeSpecName: "kube-api-access-zbctp") pod "bac1c5d6-a775-483f-ad66-20143a6b5141" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141"). InnerVolumeSpecName "kube-api-access-zbctp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.511613 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bac1c5d6-a775-483f-ad66-20143a6b5141" (UID: "bac1c5d6-a775-483f-ad66-20143a6b5141"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.539441 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:38 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:38 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:38 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.539513 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.568658 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrzt2"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.602151 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbctp\" (UniqueName: \"kubernetes.io/projected/bac1c5d6-a775-483f-ad66-20143a6b5141-kube-api-access-zbctp\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.602668 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac1c5d6-a775-483f-ad66-20143a6b5141-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.602680 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.602690 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac1c5d6-a775-483f-ad66-20143a6b5141-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:38 crc kubenswrapper[4843]: W0314 09:12:38.643355 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf36123e7_f666_4382_8324_252e94a27884.slice/crio-3e6567d361d83eb85b5c195629f4136eef79fde8b6b4fbbebc8bb7c975bfaf9f WatchSource:0}: Error finding container 3e6567d361d83eb85b5c195629f4136eef79fde8b6b4fbbebc8bb7c975bfaf9f: Status 404 returned error can't find the container with id 3e6567d361d83eb85b5c195629f4136eef79fde8b6b4fbbebc8bb7c975bfaf9f Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.803578 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerStarted","Data":"3e6567d361d83eb85b5c195629f4136eef79fde8b6b4fbbebc8bb7c975bfaf9f"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.806352 4843 generic.go:334] "Generic (PLEG): container finished" podID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerID="12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45" exitCode=0 Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.806395 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx9cb" event={"ID":"3b7c5681-b6b0-4bfd-97d8-741096bb89d0","Type":"ContainerDied","Data":"12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.806448 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx9cb" event={"ID":"3b7c5681-b6b0-4bfd-97d8-741096bb89d0","Type":"ContainerStarted","Data":"02ca0dd7fb363d2a992b7a208943e194d21266d94339fdfec2781ee4931410b8"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.808239 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.808325 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" event={"ID":"cf1e453b-16e7-4413-97f2-ebf025a6f6d0","Type":"ContainerStarted","Data":"b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.808361 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" event={"ID":"cf1e453b-16e7-4413-97f2-ebf025a6f6d0","Type":"ContainerStarted","Data":"21e49ab2218cbcc757b33dc9fd636d66c2d781eeceaab0adcc5b1d5740e9273c"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.808416 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.809950 4843 generic.go:334] "Generic (PLEG): container finished" podID="bac1c5d6-a775-483f-ad66-20143a6b5141" containerID="1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e" exitCode=0 Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.810024 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" event={"ID":"bac1c5d6-a775-483f-ad66-20143a6b5141","Type":"ContainerDied","Data":"1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.810043 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" event={"ID":"bac1c5d6-a775-483f-ad66-20143a6b5141","Type":"ContainerDied","Data":"3866973b24d16455e5b78997a668e196ffdae40c53c864addb50d7ffc80341d8"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.810060 4843 scope.go:117] "RemoveContainer" containerID="1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.810149 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.824599 4843 generic.go:334] "Generic (PLEG): container finished" podID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerID="2a209684e8e5c6b45db759e4ffa84a740aabdbcf8d9effb6c818e1694d55efc0" exitCode=0 Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.824683 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzj4r" event={"ID":"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f","Type":"ContainerDied","Data":"2a209684e8e5c6b45db759e4ffa84a740aabdbcf8d9effb6c818e1694d55efc0"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.824717 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzj4r" event={"ID":"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f","Type":"ContainerStarted","Data":"72e753957c21bdc806b07b295fe82dc111a25d82d3e3235682af56fcd1d612dd"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.836331 4843 scope.go:117] "RemoveContainer" containerID="1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e" Mar 14 09:12:38 crc kubenswrapper[4843]: E0314 09:12:38.836789 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e\": container with ID starting with 1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e not found: ID does not exist" containerID="1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.836822 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e"} err="failed to get container status \"1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e\": rpc error: code = NotFound desc = could not find container \"1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e\": container with ID starting with 1ac68c915e3f8ef6d9e16a29015b812b56465d1e53423e7b6f1658499a15d08e not found: ID does not exist" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.837365 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerID="bc486ebf3271c5a67c24e45698b89cc184cfd9d386e70c7553e094393635cfeb" exitCode=0 Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.838052 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mss9d" event={"ID":"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb","Type":"ContainerDied","Data":"bc486ebf3271c5a67c24e45698b89cc184cfd9d386e70c7553e094393635cfeb"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.838118 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mss9d" event={"ID":"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb","Type":"ContainerStarted","Data":"7ef4eb91812527ad7654bccd71811449a0a07f27f2faf65e8a1f28b792f0b475"} Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.852422 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" podStartSLOduration=52.85240431 podStartE2EDuration="52.85240431s" podCreationTimestamp="2026-03-14 09:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:38.849616376 +0000 UTC m=+126.162227504" watchObservedRunningTime="2026-03-14 09:12:38.85240431 +0000 UTC m=+126.165015438" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.893724 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.896755 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4nzm6"] Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.995956 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57866b686d-s8m62"] Mar 14 09:12:38 crc kubenswrapper[4843]: E0314 09:12:38.996259 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac1c5d6-a775-483f-ad66-20143a6b5141" containerName="route-controller-manager" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.996312 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac1c5d6-a775-483f-ad66-20143a6b5141" containerName="route-controller-manager" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.996430 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac1c5d6-a775-483f-ad66-20143a6b5141" containerName="route-controller-manager" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.996967 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.999623 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 14 09:12:38 crc kubenswrapper[4843]: I0314 09:12:38.999702 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.000363 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.000581 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.001287 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.001292 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.017152 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.017610 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.017820 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.017900 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.018134 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.018240 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.018383 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.018529 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57866b686d-s8m62"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.018659 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.021288 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.021748 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.110300 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c222d32-5588-40c0-a714-dc50cd379a67-serving-cert\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.110852 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j27r4\" (UniqueName: \"kubernetes.io/projected/b32180a9-0ae7-4384-876a-345dc2b6ec27-kube-api-access-j27r4\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.110887 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-client-ca\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.110937 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-client-ca\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.110980 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-proxy-ca-bundles\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.111002 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-config\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.111035 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt465\" (UniqueName: \"kubernetes.io/projected/4c222d32-5588-40c0-a714-dc50cd379a67-kube-api-access-lt465\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.111079 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.111114 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.111148 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b32180a9-0ae7-4384-876a-345dc2b6ec27-serving-cert\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.111175 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-config\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.113298 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.113586 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.123813 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.130590 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212642 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212712 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b32180a9-0ae7-4384-876a-345dc2b6ec27-serving-cert\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212755 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-config\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212792 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c222d32-5588-40c0-a714-dc50cd379a67-serving-cert\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212828 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j27r4\" (UniqueName: \"kubernetes.io/projected/b32180a9-0ae7-4384-876a-345dc2b6ec27-kube-api-access-j27r4\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212867 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-client-ca\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212907 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.212950 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-client-ca\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.213003 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-proxy-ca-bundles\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.213027 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-config\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.213060 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt465\" (UniqueName: \"kubernetes.io/projected/4c222d32-5588-40c0-a714-dc50cd379a67-kube-api-access-lt465\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.214446 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-client-ca\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.214559 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-config\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.214890 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-client-ca\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.214948 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.215399 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-proxy-ca-bundles\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.215737 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-config\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.217375 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c222d32-5588-40c0-a714-dc50cd379a67-serving-cert\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.217650 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b32180a9-0ae7-4384-876a-345dc2b6ec27-serving-cert\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.225069 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.236425 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.238623 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.239684 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt465\" (UniqueName: \"kubernetes.io/projected/4c222d32-5588-40c0-a714-dc50cd379a67-kube-api-access-lt465\") pod \"controller-manager-57866b686d-s8m62\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.240744 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j27r4\" (UniqueName: \"kubernetes.io/projected/b32180a9-0ae7-4384-876a-345dc2b6ec27-kube-api-access-j27r4\") pod \"route-controller-manager-cd676d8c9-sv9b6\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.257414 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.267741 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.283847 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.312108 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.330528 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.344657 4843 patch_prober.go:28] interesting pod/downloads-7954f5f757-tr9zk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.344668 4843 patch_prober.go:28] interesting pod/downloads-7954f5f757-tr9zk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.344732 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tr9zk" podUID="570cf110-147b-482b-afe9-2b0ca133b6dc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.344765 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tr9zk" podUID="570cf110-147b-482b-afe9-2b0ca133b6dc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.350792 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1074b905-740e-4fc3-919b-e20309a9098b" path="/var/lib/kubelet/pods/1074b905-740e-4fc3-919b-e20309a9098b/volumes" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.351706 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.352407 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bac1c5d6-a775-483f-ad66-20143a6b5141" path="/var/lib/kubelet/pods/bac1c5d6-a775-483f-ad66-20143a6b5141/volumes" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.353695 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mmpq6"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.355089 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.359698 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.367907 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmpq6"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.402233 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.403206 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.404526 4843 patch_prober.go:28] interesting pod/console-f9d7485db-2j8wl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.404563 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2j8wl" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.517027 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-utilities\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.517079 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-catalog-content\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.518170 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmtt4\" (UniqueName: \"kubernetes.io/projected/8220c353-b74c-4703-ac79-af96379e89fe-kube-api-access-gmtt4\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.535565 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.543191 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:39 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:39 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:39 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.543242 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.574337 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.575000 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.576663 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.577656 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.587318 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.620002 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-utilities\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.620045 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-catalog-content\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.620076 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmtt4\" (UniqueName: \"kubernetes.io/projected/8220c353-b74c-4703-ac79-af96379e89fe-kube-api-access-gmtt4\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.620682 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-utilities\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.620795 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-catalog-content\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.648656 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmtt4\" (UniqueName: \"kubernetes.io/projected/8220c353-b74c-4703-ac79-af96379e89fe-kube-api-access-gmtt4\") pod \"redhat-marketplace-mmpq6\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.686183 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.721753 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.721826 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.757338 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xg5hq"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.759628 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xg5hq"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.759730 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:39 crc kubenswrapper[4843]: W0314 09:12:39.803929 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-92a4fda622be041add5993cceee12f6cf4f189499a4209d9b353e3a5fe25f0fc WatchSource:0}: Error finding container 92a4fda622be041add5993cceee12f6cf4f189499a4209d9b353e3a5fe25f0fc: Status 404 returned error can't find the container with id 92a4fda622be041add5993cceee12f6cf4f189499a4209d9b353e3a5fe25f0fc Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.824994 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.825076 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.825150 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.829343 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57866b686d-s8m62"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.830603 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6"] Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.846967 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: W0314 09:12:39.861638 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb32180a9_0ae7_4384_876a_345dc2b6ec27.slice/crio-99e27590273574986f681c43482b705deb6172d34082ffeae0a95b13cb846026 WatchSource:0}: Error finding container 99e27590273574986f681c43482b705deb6172d34082ffeae0a95b13cb846026: Status 404 returned error can't find the container with id 99e27590273574986f681c43482b705deb6172d34082ffeae0a95b13cb846026 Mar 14 09:12:39 crc kubenswrapper[4843]: W0314 09:12:39.861977 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c222d32_5588_40c0_a714_dc50cd379a67.slice/crio-cb90bc8d7bf1dd4964cd9d6955e01335ad90d48e184605a99e97a45b181a22a6 WatchSource:0}: Error finding container cb90bc8d7bf1dd4964cd9d6955e01335ad90d48e184605a99e97a45b181a22a6: Status 404 returned error can't find the container with id cb90bc8d7bf1dd4964cd9d6955e01335ad90d48e184605a99e97a45b181a22a6 Mar 14 09:12:39 crc kubenswrapper[4843]: W0314 09:12:39.866065 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-d859d3cfc7f4de921645e1e26107147e3bfb5b22c835120b9146db6ed407d2ed WatchSource:0}: Error finding container d859d3cfc7f4de921645e1e26107147e3bfb5b22c835120b9146db6ed407d2ed: Status 404 returned error can't find the container with id d859d3cfc7f4de921645e1e26107147e3bfb5b22c835120b9146db6ed407d2ed Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.869200 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"92a4fda622be041add5993cceee12f6cf4f189499a4209d9b353e3a5fe25f0fc"} Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.871300 4843 generic.go:334] "Generic (PLEG): container finished" podID="f36123e7-f666-4382-8324-252e94a27884" containerID="39fbee9e3534a3ba4a6ba3d977073f5515ed8b7c35c85a9aab284b1d5ffd87a3" exitCode=0 Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.871351 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerDied","Data":"39fbee9e3534a3ba4a6ba3d977073f5515ed8b7c35c85a9aab284b1d5ffd87a3"} Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.882352 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c66cb7bc273b1ab7fede2046a0719c444fc8a918713473cd6dd6a66b293fbae1"} Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.882391 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8c47ac25807113a8fb7474e2f531ec3b0526a4624a2fe69efd02da23711e3f3f"} Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.882888 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.910938 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.936665 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-utilities\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.936718 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbvwl\" (UniqueName: \"kubernetes.io/projected/59b75c6f-f60e-4ed2-8dec-d61a29121a63-kube-api-access-hbvwl\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.936960 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-catalog-content\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:39 crc kubenswrapper[4843]: I0314 09:12:39.980746 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmpq6"] Mar 14 09:12:39 crc kubenswrapper[4843]: W0314 09:12:39.990058 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8220c353_b74c_4703_ac79_af96379e89fe.slice/crio-f8b73f6dcdbb225316a081ed59715469f3075e11b49b7eb310bc589582ca274c WatchSource:0}: Error finding container f8b73f6dcdbb225316a081ed59715469f3075e11b49b7eb310bc589582ca274c: Status 404 returned error can't find the container with id f8b73f6dcdbb225316a081ed59715469f3075e11b49b7eb310bc589582ca274c Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.037749 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-catalog-content\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.037788 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-utilities\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.037803 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbvwl\" (UniqueName: \"kubernetes.io/projected/59b75c6f-f60e-4ed2-8dec-d61a29121a63-kube-api-access-hbvwl\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.039962 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-catalog-content\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.040287 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-utilities\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.066530 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbvwl\" (UniqueName: \"kubernetes.io/projected/59b75c6f-f60e-4ed2-8dec-d61a29121a63-kube-api-access-hbvwl\") pod \"redhat-marketplace-xg5hq\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.081313 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.516447 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.518767 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.545064 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:40 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:40 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:40 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.545124 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.546446 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.580370 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xg5hq"] Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.583075 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.745441 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jfzjg"] Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.746949 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: E0314 09:12:40.747400 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.749615 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 14 09:12:40 crc kubenswrapper[4843]: E0314 09:12:40.750011 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:12:40 crc kubenswrapper[4843]: E0314 09:12:40.761500 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:12:40 crc kubenswrapper[4843]: E0314 09:12:40.761577 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.763571 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jfzjg"] Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.849057 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5z9q\" (UniqueName: \"kubernetes.io/projected/3568e895-1a84-487a-9042-376d2bb9e28f-kube-api-access-t5z9q\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.849514 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-utilities\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.849561 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-catalog-content\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.898430 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"92cf5ee9793d28e42602499a128f77fe731053bee965726281761db60b66963c"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.898479 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d859d3cfc7f4de921645e1e26107147e3bfb5b22c835120b9146db6ed407d2ed"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.902293 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" event={"ID":"4c222d32-5588-40c0-a714-dc50cd379a67","Type":"ContainerStarted","Data":"4e4e674e19d11abec233234b6607e6148d0893938b8607b0bf4ef79184d18964"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.902323 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" event={"ID":"4c222d32-5588-40c0-a714-dc50cd379a67","Type":"ContainerStarted","Data":"cb90bc8d7bf1dd4964cd9d6955e01335ad90d48e184605a99e97a45b181a22a6"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.902977 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.903971 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xg5hq" event={"ID":"59b75c6f-f60e-4ed2-8dec-d61a29121a63","Type":"ContainerStarted","Data":"92ba406da930481e2a77149d259e1eac1d46876f8c7f706f28095abc37ae93df"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.906423 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c6fe3ef4-be53-4048-ad60-5bdf196cf10e","Type":"ContainerStarted","Data":"4d5f07ac339802fd9e6fae8a29cab6ea00f7265f7337d8ad258e958a2e2e089f"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.909773 4843 generic.go:334] "Generic (PLEG): container finished" podID="8220c353-b74c-4703-ac79-af96379e89fe" containerID="035fd8923dba17fd8d08379cd43918a0f3ac4aa51fc14f6e35af74b8e9d9978a" exitCode=0 Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.909944 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmpq6" event={"ID":"8220c353-b74c-4703-ac79-af96379e89fe","Type":"ContainerDied","Data":"035fd8923dba17fd8d08379cd43918a0f3ac4aa51fc14f6e35af74b8e9d9978a"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.909968 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmpq6" event={"ID":"8220c353-b74c-4703-ac79-af96379e89fe","Type":"ContainerStarted","Data":"f8b73f6dcdbb225316a081ed59715469f3075e11b49b7eb310bc589582ca274c"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.912317 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.918437 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" event={"ID":"b32180a9-0ae7-4384-876a-345dc2b6ec27","Type":"ContainerStarted","Data":"f0f2a64c4aedcac2b813fb694ba20ea7602d5a790225732439d0853358d5ddb3"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.918475 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" event={"ID":"b32180a9-0ae7-4384-876a-345dc2b6ec27","Type":"ContainerStarted","Data":"99e27590273574986f681c43482b705deb6172d34082ffeae0a95b13cb846026"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.919266 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.924915 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"04715064088b383b898a1fed4f05fcd26799db1ddc714bf845159e8485b8edd7"} Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.927577 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.936735 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-n64hl" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.949867 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" podStartSLOduration=4.949847255 podStartE2EDuration="4.949847255s" podCreationTimestamp="2026-03-14 09:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:40.949234637 +0000 UTC m=+128.261845765" watchObservedRunningTime="2026-03-14 09:12:40.949847255 +0000 UTC m=+128.262458383" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.953834 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5z9q\" (UniqueName: \"kubernetes.io/projected/3568e895-1a84-487a-9042-376d2bb9e28f-kube-api-access-t5z9q\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.955327 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-utilities\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.955484 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-catalog-content\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.956087 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-catalog-content\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.957603 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-utilities\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.962262 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.962309 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.968802 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.978377 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5z9q\" (UniqueName: \"kubernetes.io/projected/3568e895-1a84-487a-9042-376d2bb9e28f-kube-api-access-t5z9q\") pod \"redhat-operators-jfzjg\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.981841 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" podStartSLOduration=3.9818255000000002 podStartE2EDuration="3.9818255s" podCreationTimestamp="2026-03-14 09:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:40.966934261 +0000 UTC m=+128.279545389" watchObservedRunningTime="2026-03-14 09:12:40.9818255 +0000 UTC m=+128.294436628" Mar 14 09:12:40 crc kubenswrapper[4843]: I0314 09:12:40.999786 4843 ???:1] "http: TLS handshake error from 192.168.126.11:43232: no serving certificate available for the kubelet" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.106824 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.152034 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9lwgw"] Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.167177 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9lwgw"] Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.168938 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.263647 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-utilities\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.263703 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-catalog-content\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.263733 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsscf\" (UniqueName: \"kubernetes.io/projected/92538af0-1d79-4c19-b7c2-f1c1b89595a3-kube-api-access-tsscf\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.365092 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-utilities\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.365138 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-catalog-content\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.365176 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsscf\" (UniqueName: \"kubernetes.io/projected/92538af0-1d79-4c19-b7c2-f1c1b89595a3-kube-api-access-tsscf\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.365702 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-utilities\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.365904 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-catalog-content\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.386353 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsscf\" (UniqueName: \"kubernetes.io/projected/92538af0-1d79-4c19-b7c2-f1c1b89595a3-kube-api-access-tsscf\") pod \"redhat-operators-9lwgw\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.425791 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.427098 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.430024 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.430042 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.436465 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.497465 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.544567 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:41 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:41 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:41 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.544923 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.568540 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.568607 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.632513 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jfzjg"] Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.670155 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.670265 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.670348 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.686467 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.748977 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.933047 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerStarted","Data":"58ee98669e99b6063d4a255923e898688b4bb00a48cbdd166a48c5976a1ea9f0"} Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.935096 4843 generic.go:334] "Generic (PLEG): container finished" podID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerID="faf12fe2b9e8ce822d8b8317c6c5779ec87701f1200c3723ec65e3ac3f79ad88" exitCode=0 Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.935168 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xg5hq" event={"ID":"59b75c6f-f60e-4ed2-8dec-d61a29121a63","Type":"ContainerDied","Data":"faf12fe2b9e8ce822d8b8317c6c5779ec87701f1200c3723ec65e3ac3f79ad88"} Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.937557 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c6fe3ef4-be53-4048-ad60-5bdf196cf10e","Type":"ContainerStarted","Data":"96e5c35dc98a8411a106783889700fba78888392f285f72c9a1811a72a88810f"} Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.943831 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4ftzb" Mar 14 09:12:41 crc kubenswrapper[4843]: I0314 09:12:41.978215 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9lwgw"] Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.097123 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.544369 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:42 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:42 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:42 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.545110 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.909881 4843 ???:1] "http: TLS handshake error from 192.168.126.11:43242: no serving certificate available for the kubelet" Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.949693 4843 generic.go:334] "Generic (PLEG): container finished" podID="3568e895-1a84-487a-9042-376d2bb9e28f" containerID="ca20868760ef13e0cc0ebee4f5469dd4fa4773c40ffc5c993defc2b36e97e34d" exitCode=0 Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.949762 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerDied","Data":"ca20868760ef13e0cc0ebee4f5469dd4fa4773c40ffc5c993defc2b36e97e34d"} Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.955208 4843 generic.go:334] "Generic (PLEG): container finished" podID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerID="6a78abfcee61c1388010cca35e259c2321f39887f19380ae91695dcc2822023d" exitCode=0 Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.955288 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerDied","Data":"6a78abfcee61c1388010cca35e259c2321f39887f19380ae91695dcc2822023d"} Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.955319 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerStarted","Data":"fc481e852e9ff72029931ba9c71dadf2f199c6ac34e743ab01c566463da2ec43"} Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.970200 4843 generic.go:334] "Generic (PLEG): container finished" podID="c6fe3ef4-be53-4048-ad60-5bdf196cf10e" containerID="96e5c35dc98a8411a106783889700fba78888392f285f72c9a1811a72a88810f" exitCode=0 Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.970566 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c6fe3ef4-be53-4048-ad60-5bdf196cf10e","Type":"ContainerDied","Data":"96e5c35dc98a8411a106783889700fba78888392f285f72c9a1811a72a88810f"} Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.985565 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ad8ab5cf-bf5d-405b-a2ff-d2d394731263","Type":"ContainerStarted","Data":"070d6b66a8d6453d5afcda52cb6cafa6a51220965cfbd293eced7d784e767196"} Mar 14 09:12:42 crc kubenswrapper[4843]: I0314 09:12:42.985604 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ad8ab5cf-bf5d-405b-a2ff-d2d394731263","Type":"ContainerStarted","Data":"95b4162ab4fefca06d0aa6ebe0d2f3157316998e58c416ea5f4fc2e7f0c92e73"} Mar 14 09:12:43 crc kubenswrapper[4843]: I0314 09:12:43.047963 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.04794391 podStartE2EDuration="2.04794391s" podCreationTimestamp="2026-03-14 09:12:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:43.045477896 +0000 UTC m=+130.358089024" watchObservedRunningTime="2026-03-14 09:12:43.04794391 +0000 UTC m=+130.360555038" Mar 14 09:12:43 crc kubenswrapper[4843]: I0314 09:12:43.538816 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:43 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:43 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:43 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:43 crc kubenswrapper[4843]: I0314 09:12:43.538876 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.016711 4843 generic.go:334] "Generic (PLEG): container finished" podID="ad8ab5cf-bf5d-405b-a2ff-d2d394731263" containerID="070d6b66a8d6453d5afcda52cb6cafa6a51220965cfbd293eced7d784e767196" exitCode=0 Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.017648 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ad8ab5cf-bf5d-405b-a2ff-d2d394731263","Type":"ContainerDied","Data":"070d6b66a8d6453d5afcda52cb6cafa6a51220965cfbd293eced7d784e767196"} Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.331649 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.480602 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kubelet-dir\") pod \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.480706 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c6fe3ef4-be53-4048-ad60-5bdf196cf10e" (UID: "c6fe3ef4-be53-4048-ad60-5bdf196cf10e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.480919 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kube-api-access\") pod \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\" (UID: \"c6fe3ef4-be53-4048-ad60-5bdf196cf10e\") " Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.481181 4843 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.488495 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c6fe3ef4-be53-4048-ad60-5bdf196cf10e" (UID: "c6fe3ef4-be53-4048-ad60-5bdf196cf10e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.540652 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:44 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:44 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:44 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.541058 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:44 crc kubenswrapper[4843]: I0314 09:12:44.582207 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6fe3ef4-be53-4048-ad60-5bdf196cf10e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:12:45 crc kubenswrapper[4843]: I0314 09:12:45.038263 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 14 09:12:45 crc kubenswrapper[4843]: I0314 09:12:45.038397 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c6fe3ef4-be53-4048-ad60-5bdf196cf10e","Type":"ContainerDied","Data":"4d5f07ac339802fd9e6fae8a29cab6ea00f7265f7337d8ad258e958a2e2e089f"} Mar 14 09:12:45 crc kubenswrapper[4843]: I0314 09:12:45.038554 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d5f07ac339802fd9e6fae8a29cab6ea00f7265f7337d8ad258e958a2e2e089f" Mar 14 09:12:45 crc kubenswrapper[4843]: I0314 09:12:45.549824 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:45 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:45 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:45 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:45 crc kubenswrapper[4843]: I0314 09:12:45.550424 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:45 crc kubenswrapper[4843]: I0314 09:12:45.630164 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:12:46 crc kubenswrapper[4843]: I0314 09:12:46.045073 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-mr75n" Mar 14 09:12:46 crc kubenswrapper[4843]: I0314 09:12:46.158051 4843 ???:1] "http: TLS handshake error from 192.168.126.11:43256: no serving certificate available for the kubelet" Mar 14 09:12:46 crc kubenswrapper[4843]: I0314 09:12:46.543809 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:46 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:46 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:46 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:46 crc kubenswrapper[4843]: I0314 09:12:46.543874 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:47 crc kubenswrapper[4843]: I0314 09:12:47.363075 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 14 09:12:47 crc kubenswrapper[4843]: I0314 09:12:47.540174 4843 patch_prober.go:28] interesting pod/router-default-5444994796-pwxsw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 14 09:12:47 crc kubenswrapper[4843]: [-]has-synced failed: reason withheld Mar 14 09:12:47 crc kubenswrapper[4843]: [+]process-running ok Mar 14 09:12:47 crc kubenswrapper[4843]: healthz check failed Mar 14 09:12:47 crc kubenswrapper[4843]: I0314 09:12:47.540249 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pwxsw" podUID="fb7125f9-9cca-41e8-9cb0-af829e1b7287" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 14 09:12:48 crc kubenswrapper[4843]: I0314 09:12:48.544075 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:48 crc kubenswrapper[4843]: I0314 09:12:48.548619 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-pwxsw" Mar 14 09:12:48 crc kubenswrapper[4843]: I0314 09:12:48.567194 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.5671764929999998 podStartE2EDuration="1.567176493s" podCreationTimestamp="2026-03-14 09:12:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:48.564677073 +0000 UTC m=+135.877288201" watchObservedRunningTime="2026-03-14 09:12:48.567176493 +0000 UTC m=+135.879787621" Mar 14 09:12:49 crc kubenswrapper[4843]: I0314 09:12:49.372618 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-tr9zk" Mar 14 09:12:49 crc kubenswrapper[4843]: I0314 09:12:49.403193 4843 patch_prober.go:28] interesting pod/console-f9d7485db-2j8wl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 14 09:12:49 crc kubenswrapper[4843]: I0314 09:12:49.403290 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2j8wl" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 14 09:12:50 crc kubenswrapper[4843]: I0314 09:12:50.306142 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:12:50 crc kubenswrapper[4843]: I0314 09:12:50.353315 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 14 09:12:50 crc kubenswrapper[4843]: E0314 09:12:50.758453 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:12:50 crc kubenswrapper[4843]: E0314 09:12:50.760389 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:12:50 crc kubenswrapper[4843]: E0314 09:12:50.763475 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:12:50 crc kubenswrapper[4843]: E0314 09:12:50.763520 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:12:53 crc kubenswrapper[4843]: I0314 09:12:53.386571 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=3.386551901 podStartE2EDuration="3.386551901s" podCreationTimestamp="2026-03-14 09:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:12:53.384061821 +0000 UTC m=+140.696672979" watchObservedRunningTime="2026-03-14 09:12:53.386551901 +0000 UTC m=+140.699163029" Mar 14 09:12:55 crc kubenswrapper[4843]: I0314 09:12:55.693213 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57866b686d-s8m62"] Mar 14 09:12:55 crc kubenswrapper[4843]: I0314 09:12:55.694077 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" containerName="controller-manager" containerID="cri-o://4e4e674e19d11abec233234b6607e6148d0893938b8607b0bf4ef79184d18964" gracePeriod=30 Mar 14 09:12:55 crc kubenswrapper[4843]: I0314 09:12:55.715655 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6"] Mar 14 09:12:55 crc kubenswrapper[4843]: I0314 09:12:55.715902 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerName="route-controller-manager" containerID="cri-o://f0f2a64c4aedcac2b813fb694ba20ea7602d5a790225732439d0853358d5ddb3" gracePeriod=30 Mar 14 09:12:57 crc kubenswrapper[4843]: I0314 09:12:57.217408 4843 generic.go:334] "Generic (PLEG): container finished" podID="4c222d32-5588-40c0-a714-dc50cd379a67" containerID="4e4e674e19d11abec233234b6607e6148d0893938b8607b0bf4ef79184d18964" exitCode=0 Mar 14 09:12:57 crc kubenswrapper[4843]: I0314 09:12:57.217545 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" event={"ID":"4c222d32-5588-40c0-a714-dc50cd379a67","Type":"ContainerDied","Data":"4e4e674e19d11abec233234b6607e6148d0893938b8607b0bf4ef79184d18964"} Mar 14 09:12:57 crc kubenswrapper[4843]: I0314 09:12:57.220203 4843 generic.go:334] "Generic (PLEG): container finished" podID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerID="f0f2a64c4aedcac2b813fb694ba20ea7602d5a790225732439d0853358d5ddb3" exitCode=0 Mar 14 09:12:57 crc kubenswrapper[4843]: I0314 09:12:57.220245 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" event={"ID":"b32180a9-0ae7-4384-876a-345dc2b6ec27","Type":"ContainerDied","Data":"f0f2a64c4aedcac2b813fb694ba20ea7602d5a790225732439d0853358d5ddb3"} Mar 14 09:12:57 crc kubenswrapper[4843]: I0314 09:12:57.784063 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:12:59 crc kubenswrapper[4843]: I0314 09:12:59.313473 4843 patch_prober.go:28] interesting pod/controller-manager-57866b686d-s8m62 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Mar 14 09:12:59 crc kubenswrapper[4843]: I0314 09:12:59.313561 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Mar 14 09:12:59 crc kubenswrapper[4843]: I0314 09:12:59.331969 4843 patch_prober.go:28] interesting pod/route-controller-manager-cd676d8c9-sv9b6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" start-of-body= Mar 14 09:12:59 crc kubenswrapper[4843]: I0314 09:12:59.332097 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" Mar 14 09:12:59 crc kubenswrapper[4843]: I0314 09:12:59.406162 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:12:59 crc kubenswrapper[4843]: I0314 09:12:59.409382 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:13:00 crc kubenswrapper[4843]: E0314 09:13:00.742453 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:13:00 crc kubenswrapper[4843]: E0314 09:13:00.745121 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:13:00 crc kubenswrapper[4843]: E0314 09:13:00.746729 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:13:00 crc kubenswrapper[4843]: E0314 09:13:00.746781 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:13:06 crc kubenswrapper[4843]: I0314 09:13:06.273123 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-mrmxq_94325ed6-8249-4f3a-98fc-fc507ab5ec18/kube-multus-additional-cni-plugins/0.log" Mar 14 09:13:06 crc kubenswrapper[4843]: I0314 09:13:06.273937 4843 generic.go:334] "Generic (PLEG): container finished" podID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" exitCode=137 Mar 14 09:13:06 crc kubenswrapper[4843]: I0314 09:13:06.273994 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" event={"ID":"94325ed6-8249-4f3a-98fc-fc507ab5ec18","Type":"ContainerDied","Data":"e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258"} Mar 14 09:13:06 crc kubenswrapper[4843]: I0314 09:13:06.674213 4843 ???:1] "http: TLS handshake error from 192.168.126.11:50860: no serving certificate available for the kubelet" Mar 14 09:13:07 crc kubenswrapper[4843]: I0314 09:13:07.929653 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.106181 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kube-api-access\") pod \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.106403 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kubelet-dir\") pod \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\" (UID: \"ad8ab5cf-bf5d-405b-a2ff-d2d394731263\") " Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.106496 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ad8ab5cf-bf5d-405b-a2ff-d2d394731263" (UID: "ad8ab5cf-bf5d-405b-a2ff-d2d394731263"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.106868 4843 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.113524 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ad8ab5cf-bf5d-405b-a2ff-d2d394731263" (UID: "ad8ab5cf-bf5d-405b-a2ff-d2d394731263"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.209592 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad8ab5cf-bf5d-405b-a2ff-d2d394731263-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.291077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ad8ab5cf-bf5d-405b-a2ff-d2d394731263","Type":"ContainerDied","Data":"95b4162ab4fefca06d0aa6ebe0d2f3157316998e58c416ea5f4fc2e7f0c92e73"} Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.291169 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95b4162ab4fefca06d0aa6ebe0d2f3157316998e58c416ea5f4fc2e7f0c92e73" Mar 14 09:13:08 crc kubenswrapper[4843]: I0314 09:13:08.291172 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 14 09:13:09 crc kubenswrapper[4843]: I0314 09:13:09.262354 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 14 09:13:09 crc kubenswrapper[4843]: I0314 09:13:09.957882 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hc4vf" Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.293826 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.294027 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xhsjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kzj4r_openshift-marketplace(076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.295250 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kzj4r" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" Mar 14 09:13:10 crc kubenswrapper[4843]: I0314 09:13:10.313141 4843 patch_prober.go:28] interesting pod/controller-manager-57866b686d-s8m62 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": context deadline exceeded" start-of-body= Mar 14 09:13:10 crc kubenswrapper[4843]: I0314 09:13:10.313210 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": context deadline exceeded" Mar 14 09:13:10 crc kubenswrapper[4843]: I0314 09:13:10.332438 4843 patch_prober.go:28] interesting pod/route-controller-manager-cd676d8c9-sv9b6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:13:10 crc kubenswrapper[4843]: I0314 09:13:10.332492 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:13:10 crc kubenswrapper[4843]: I0314 09:13:10.351859 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.740787 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258 is running failed: container process not found" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.740991 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258 is running failed: container process not found" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.741339 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258 is running failed: container process not found" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 14 09:13:10 crc kubenswrapper[4843]: E0314 09:13:10.741368 4843 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:13:13 crc kubenswrapper[4843]: I0314 09:13:13.354507 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.354489293 podStartE2EDuration="3.354489293s" podCreationTimestamp="2026-03-14 09:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:13.352252021 +0000 UTC m=+160.664863149" watchObservedRunningTime="2026-03-14 09:13:13.354489293 +0000 UTC m=+160.667100421" Mar 14 09:13:13 crc kubenswrapper[4843]: E0314 09:13:13.673675 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kzj4r" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" Mar 14 09:13:13 crc kubenswrapper[4843]: E0314 09:13:13.760620 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 14 09:13:13 crc kubenswrapper[4843]: E0314 09:13:13.760772 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tsscf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9lwgw_openshift-marketplace(92538af0-1d79-4c19-b7c2-f1c1b89595a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:13 crc kubenswrapper[4843]: E0314 09:13:13.761962 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9lwgw" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.821423 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 14 09:13:14 crc kubenswrapper[4843]: E0314 09:13:14.822056 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fe3ef4-be53-4048-ad60-5bdf196cf10e" containerName="pruner" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.822070 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fe3ef4-be53-4048-ad60-5bdf196cf10e" containerName="pruner" Mar 14 09:13:14 crc kubenswrapper[4843]: E0314 09:13:14.822078 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8ab5cf-bf5d-405b-a2ff-d2d394731263" containerName="pruner" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.822085 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8ab5cf-bf5d-405b-a2ff-d2d394731263" containerName="pruner" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.822189 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad8ab5cf-bf5d-405b-a2ff-d2d394731263" containerName="pruner" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.822205 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6fe3ef4-be53-4048-ad60-5bdf196cf10e" containerName="pruner" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.822608 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.828611 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.828982 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 14 09:13:14 crc kubenswrapper[4843]: I0314 09:13:14.832427 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.003754 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/684ccdc3-1bcb-4227-8679-b9d31a21198c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.004176 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/684ccdc3-1bcb-4227-8679-b9d31a21198c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.105078 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/684ccdc3-1bcb-4227-8679-b9d31a21198c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.105201 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/684ccdc3-1bcb-4227-8679-b9d31a21198c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.105199 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/684ccdc3-1bcb-4227-8679-b9d31a21198c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.126556 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/684ccdc3-1bcb-4227-8679-b9d31a21198c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: I0314 09:13:15.159145 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:15 crc kubenswrapper[4843]: E0314 09:13:15.665690 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9lwgw" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" Mar 14 09:13:15 crc kubenswrapper[4843]: E0314 09:13:15.949981 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 14 09:13:15 crc kubenswrapper[4843]: E0314 09:13:15.950556 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7lhj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zx9cb_openshift-marketplace(3b7c5681-b6b0-4bfd-97d8-741096bb89d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:15 crc kubenswrapper[4843]: E0314 09:13:15.952108 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zx9cb" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" Mar 14 09:13:16 crc kubenswrapper[4843]: E0314 09:13:16.171591 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 14 09:13:16 crc kubenswrapper[4843]: E0314 09:13:16.171801 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-krq99,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vrzt2_openshift-marketplace(f36123e7-f666-4382-8324-252e94a27884): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:16 crc kubenswrapper[4843]: E0314 09:13:16.173031 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vrzt2" podUID="f36123e7-f666-4382-8324-252e94a27884" Mar 14 09:13:16 crc kubenswrapper[4843]: E0314 09:13:16.189528 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 14 09:13:16 crc kubenswrapper[4843]: E0314 09:13:16.189710 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g98sp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mss9d_openshift-marketplace(0c1ae1f4-158a-4e77-ae1f-1f5db661bafb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:16 crc kubenswrapper[4843]: E0314 09:13:16.190895 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mss9d" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.229464 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mss9d" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.229483 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vrzt2" podUID="f36123e7-f666-4382-8324-252e94a27884" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.229483 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zx9cb" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.298047 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.298244 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hbvwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xg5hq_openshift-marketplace(59b75c6f-f60e-4ed2-8dec-d61a29121a63): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.299400 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xg5hq" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.315046 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.320227 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-mrmxq_94325ed6-8249-4f3a-98fc-fc507ab5ec18/kube-multus-additional-cni-plugins/0.log" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.320323 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.326065 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.344786 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/94325ed6-8249-4f3a-98fc-fc507ab5ec18-ready\") pod \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.344931 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt465\" (UniqueName: \"kubernetes.io/projected/4c222d32-5588-40c0-a714-dc50cd379a67-kube-api-access-lt465\") pod \"4c222d32-5588-40c0-a714-dc50cd379a67\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.344954 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-config\") pod \"b32180a9-0ae7-4384-876a-345dc2b6ec27\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345075 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-config\") pod \"4c222d32-5588-40c0-a714-dc50cd379a67\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345101 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-proxy-ca-bundles\") pod \"4c222d32-5588-40c0-a714-dc50cd379a67\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345116 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9kfz\" (UniqueName: \"kubernetes.io/projected/94325ed6-8249-4f3a-98fc-fc507ab5ec18-kube-api-access-h9kfz\") pod \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345229 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b32180a9-0ae7-4384-876a-345dc2b6ec27-serving-cert\") pod \"b32180a9-0ae7-4384-876a-345dc2b6ec27\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345251 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j27r4\" (UniqueName: \"kubernetes.io/projected/b32180a9-0ae7-4384-876a-345dc2b6ec27-kube-api-access-j27r4\") pod \"b32180a9-0ae7-4384-876a-345dc2b6ec27\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345368 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/94325ed6-8249-4f3a-98fc-fc507ab5ec18-tuning-conf-dir\") pod \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345386 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-client-ca\") pod \"b32180a9-0ae7-4384-876a-345dc2b6ec27\" (UID: \"b32180a9-0ae7-4384-876a-345dc2b6ec27\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345401 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/94325ed6-8249-4f3a-98fc-fc507ab5ec18-cni-sysctl-allowlist\") pod \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\" (UID: \"94325ed6-8249-4f3a-98fc-fc507ab5ec18\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345419 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-client-ca\") pod \"4c222d32-5588-40c0-a714-dc50cd379a67\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345553 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c222d32-5588-40c0-a714-dc50cd379a67-serving-cert\") pod \"4c222d32-5588-40c0-a714-dc50cd379a67\" (UID: \"4c222d32-5588-40c0-a714-dc50cd379a67\") " Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.345951 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-config" (OuterVolumeSpecName: "config") pod "b32180a9-0ae7-4384-876a-345dc2b6ec27" (UID: "b32180a9-0ae7-4384-876a-345dc2b6ec27"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.346204 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-client-ca" (OuterVolumeSpecName: "client-ca") pod "b32180a9-0ae7-4384-876a-345dc2b6ec27" (UID: "b32180a9-0ae7-4384-876a-345dc2b6ec27"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.346238 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94325ed6-8249-4f3a-98fc-fc507ab5ec18-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "94325ed6-8249-4f3a-98fc-fc507ab5ec18" (UID: "94325ed6-8249-4f3a-98fc-fc507ab5ec18"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.346565 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-client-ca" (OuterVolumeSpecName: "client-ca") pod "4c222d32-5588-40c0-a714-dc50cd379a67" (UID: "4c222d32-5588-40c0-a714-dc50cd379a67"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.346704 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94325ed6-8249-4f3a-98fc-fc507ab5ec18-ready" (OuterVolumeSpecName: "ready") pod "94325ed6-8249-4f3a-98fc-fc507ab5ec18" (UID: "94325ed6-8249-4f3a-98fc-fc507ab5ec18"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.349296 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4c222d32-5588-40c0-a714-dc50cd379a67" (UID: "4c222d32-5588-40c0-a714-dc50cd379a67"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.349731 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-config" (OuterVolumeSpecName: "config") pod "4c222d32-5588-40c0-a714-dc50cd379a67" (UID: "4c222d32-5588-40c0-a714-dc50cd379a67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.352668 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.352845 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gmtt4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mmpq6_openshift-marketplace(8220c353-b74c-4703-ac79-af96379e89fe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.354578 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mmpq6" podUID="8220c353-b74c-4703-ac79-af96379e89fe" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.355398 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94325ed6-8249-4f3a-98fc-fc507ab5ec18-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "94325ed6-8249-4f3a-98fc-fc507ab5ec18" (UID: "94325ed6-8249-4f3a-98fc-fc507ab5ec18"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.359393 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b32180a9-0ae7-4384-876a-345dc2b6ec27-kube-api-access-j27r4" (OuterVolumeSpecName: "kube-api-access-j27r4") pod "b32180a9-0ae7-4384-876a-345dc2b6ec27" (UID: "b32180a9-0ae7-4384-876a-345dc2b6ec27"). InnerVolumeSpecName "kube-api-access-j27r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.359516 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b32180a9-0ae7-4384-876a-345dc2b6ec27-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b32180a9-0ae7-4384-876a-345dc2b6ec27" (UID: "b32180a9-0ae7-4384-876a-345dc2b6ec27"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.360816 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c222d32-5588-40c0-a714-dc50cd379a67-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4c222d32-5588-40c0-a714-dc50cd379a67" (UID: "4c222d32-5588-40c0-a714-dc50cd379a67"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.370463 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94325ed6-8249-4f3a-98fc-fc507ab5ec18-kube-api-access-h9kfz" (OuterVolumeSpecName: "kube-api-access-h9kfz") pod "94325ed6-8249-4f3a-98fc-fc507ab5ec18" (UID: "94325ed6-8249-4f3a-98fc-fc507ab5ec18"). InnerVolumeSpecName "kube-api-access-h9kfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.373435 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c222d32-5588-40c0-a714-dc50cd379a67-kube-api-access-lt465" (OuterVolumeSpecName: "kube-api-access-lt465") pod "4c222d32-5588-40c0-a714-dc50cd379a67" (UID: "4c222d32-5588-40c0-a714-dc50cd379a67"). InnerVolumeSpecName "kube-api-access-lt465". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.376393 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-mrmxq_94325ed6-8249-4f3a-98fc-fc507ab5ec18/kube-multus-additional-cni-plugins/0.log" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.376672 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.380113 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.382238 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.382395 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t5z9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jfzjg_openshift-marketplace(3568e895-1a84-487a-9042-376d2bb9e28f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.384884 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jfzjg" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.402649 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408536 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mrmxq" event={"ID":"94325ed6-8249-4f3a-98fc-fc507ab5ec18","Type":"ContainerDied","Data":"dd28d4d119a552603d8de09fbadb3b31ab48a3d59960c470a53c2524bfc299d6"} Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408591 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6" event={"ID":"b32180a9-0ae7-4384-876a-345dc2b6ec27","Type":"ContainerDied","Data":"99e27590273574986f681c43482b705deb6172d34082ffeae0a95b13cb846026"} Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408607 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b6856db49-hcfzr"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408684 4843 scope.go:117] "RemoveContainer" containerID="e62e45f48789578a780a0e790ec5832634ab22ff6b371b0bfacc050567941258" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.408773 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" containerName="controller-manager" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408785 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" containerName="controller-manager" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.408795 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408801 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:13:17 crc kubenswrapper[4843]: E0314 09:13:17.408816 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerName="route-controller-manager" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408822 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerName="route-controller-manager" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408918 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" containerName="kube-multus-additional-cni-plugins" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408928 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" containerName="route-controller-manager" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.408941 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" containerName="controller-manager" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.409291 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57866b686d-s8m62" event={"ID":"4c222d32-5588-40c0-a714-dc50cd379a67","Type":"ContainerDied","Data":"cb90bc8d7bf1dd4964cd9d6955e01335ad90d48e184605a99e97a45b181a22a6"} Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.409313 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b6856db49-hcfzr"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.409391 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.413835 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.415214 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.415467 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.415629 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.415881 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.416846 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.420959 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446697 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-config\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446741 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-proxy-ca-bundles\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446782 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sfwr\" (UniqueName: \"kubernetes.io/projected/4157111b-6a9f-4047-83c8-02c6de822519-kube-api-access-6sfwr\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446799 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-client-ca\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446840 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4157111b-6a9f-4047-83c8-02c6de822519-serving-cert\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446885 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt465\" (UniqueName: \"kubernetes.io/projected/4c222d32-5588-40c0-a714-dc50cd379a67-kube-api-access-lt465\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446897 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446906 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446917 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446925 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b32180a9-0ae7-4384-876a-345dc2b6ec27-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446935 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9kfz\" (UniqueName: \"kubernetes.io/projected/94325ed6-8249-4f3a-98fc-fc507ab5ec18-kube-api-access-h9kfz\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446943 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j27r4\" (UniqueName: \"kubernetes.io/projected/b32180a9-0ae7-4384-876a-345dc2b6ec27-kube-api-access-j27r4\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446951 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b32180a9-0ae7-4384-876a-345dc2b6ec27-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446960 4843 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/94325ed6-8249-4f3a-98fc-fc507ab5ec18-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446969 4843 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/94325ed6-8249-4f3a-98fc-fc507ab5ec18-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446979 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c222d32-5588-40c0-a714-dc50cd379a67-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446989 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c222d32-5588-40c0-a714-dc50cd379a67-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.446997 4843 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/94325ed6-8249-4f3a-98fc-fc507ab5ec18-ready\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.463702 4843 scope.go:117] "RemoveContainer" containerID="f0f2a64c4aedcac2b813fb694ba20ea7602d5a790225732439d0853358d5ddb3" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.470069 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mrmxq"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.475817 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mrmxq"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.482799 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.491814 4843 scope.go:117] "RemoveContainer" containerID="4e4e674e19d11abec233234b6607e6148d0893938b8607b0bf4ef79184d18964" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.502519 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd676d8c9-sv9b6"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.510722 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57866b686d-s8m62"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.514404 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-57866b686d-s8m62"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.516106 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.549130 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sfwr\" (UniqueName: \"kubernetes.io/projected/4157111b-6a9f-4047-83c8-02c6de822519-kube-api-access-6sfwr\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.549894 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-client-ca\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.549992 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4157111b-6a9f-4047-83c8-02c6de822519-serving-cert\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.550212 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-config\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.550312 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-proxy-ca-bundles\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.550656 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-client-ca\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.553859 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-proxy-ca-bundles\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.554291 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-config\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.555459 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4157111b-6a9f-4047-83c8-02c6de822519-serving-cert\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.564612 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sfwr\" (UniqueName: \"kubernetes.io/projected/4157111b-6a9f-4047-83c8-02c6de822519-kube-api-access-6sfwr\") pod \"controller-manager-6b6856db49-hcfzr\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:17 crc kubenswrapper[4843]: I0314 09:13:17.766456 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.180887 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b6856db49-hcfzr"] Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.410025 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" event={"ID":"4157111b-6a9f-4047-83c8-02c6de822519","Type":"ContainerStarted","Data":"5694aba519048bc1f73ff8c6745b47b7465f7617160983d62cf8b8f0f1c71d74"} Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.410082 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" event={"ID":"4157111b-6a9f-4047-83c8-02c6de822519","Type":"ContainerStarted","Data":"b3b29f2d3ce60d61ac7760293902286f039a22e793aa409904c76b3de3202ab3"} Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.410376 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.413297 4843 patch_prober.go:28] interesting pod/controller-manager-6b6856db49-hcfzr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.413372 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" podUID="4157111b-6a9f-4047-83c8-02c6de822519" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.414686 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"684ccdc3-1bcb-4227-8679-b9d31a21198c","Type":"ContainerStarted","Data":"1fed3525f9457a5ace28a1ff185885717cf3804794d30d1c0c5a8f4caa84b1ba"} Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.414751 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"684ccdc3-1bcb-4227-8679-b9d31a21198c","Type":"ContainerStarted","Data":"35386648194bcd292c97ec691bd21d7e1a30dee90f9631cd3e678ac55493f178"} Mar 14 09:13:18 crc kubenswrapper[4843]: E0314 09:13:18.418926 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jfzjg" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" Mar 14 09:13:18 crc kubenswrapper[4843]: E0314 09:13:18.419071 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mmpq6" podUID="8220c353-b74c-4703-ac79-af96379e89fe" Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.425993 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" podStartSLOduration=3.425962872 podStartE2EDuration="3.425962872s" podCreationTimestamp="2026-03-14 09:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:18.423331439 +0000 UTC m=+165.735942567" watchObservedRunningTime="2026-03-14 09:13:18.425962872 +0000 UTC m=+165.738574030" Mar 14 09:13:18 crc kubenswrapper[4843]: I0314 09:13:18.498077 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=4.498048662 podStartE2EDuration="4.498048662s" podCreationTimestamp="2026-03-14 09:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:18.496583021 +0000 UTC m=+165.809194159" watchObservedRunningTime="2026-03-14 09:13:18.498048662 +0000 UTC m=+165.810659810" Mar 14 09:13:19 crc kubenswrapper[4843]: I0314 09:13:19.347360 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c222d32-5588-40c0-a714-dc50cd379a67" path="/var/lib/kubelet/pods/4c222d32-5588-40c0-a714-dc50cd379a67/volumes" Mar 14 09:13:19 crc kubenswrapper[4843]: I0314 09:13:19.348386 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94325ed6-8249-4f3a-98fc-fc507ab5ec18" path="/var/lib/kubelet/pods/94325ed6-8249-4f3a-98fc-fc507ab5ec18/volumes" Mar 14 09:13:19 crc kubenswrapper[4843]: I0314 09:13:19.349036 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b32180a9-0ae7-4384-876a-345dc2b6ec27" path="/var/lib/kubelet/pods/b32180a9-0ae7-4384-876a-345dc2b6ec27/volumes" Mar 14 09:13:19 crc kubenswrapper[4843]: I0314 09:13:19.424435 4843 generic.go:334] "Generic (PLEG): container finished" podID="684ccdc3-1bcb-4227-8679-b9d31a21198c" containerID="1fed3525f9457a5ace28a1ff185885717cf3804794d30d1c0c5a8f4caa84b1ba" exitCode=0 Mar 14 09:13:19 crc kubenswrapper[4843]: I0314 09:13:19.424570 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"684ccdc3-1bcb-4227-8679-b9d31a21198c","Type":"ContainerDied","Data":"1fed3525f9457a5ace28a1ff185885717cf3804794d30d1c0c5a8f4caa84b1ba"} Mar 14 09:13:19 crc kubenswrapper[4843]: I0314 09:13:19.429598 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.024342 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn"] Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.025028 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.027246 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.027623 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.027796 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.027874 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.028146 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.028152 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.039574 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn"] Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.085444 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb01f64-2d7e-4d36-b660-8195a07ccf14-serving-cert\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.085524 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-config\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.085560 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqm4m\" (UniqueName: \"kubernetes.io/projected/deb01f64-2d7e-4d36-b660-8195a07ccf14-kube-api-access-xqm4m\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.085596 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-client-ca\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.186467 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-config\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.186556 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqm4m\" (UniqueName: \"kubernetes.io/projected/deb01f64-2d7e-4d36-b660-8195a07ccf14-kube-api-access-xqm4m\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.186603 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-client-ca\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.186643 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb01f64-2d7e-4d36-b660-8195a07ccf14-serving-cert\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.188907 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-client-ca\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.189130 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-config\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.205330 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqm4m\" (UniqueName: \"kubernetes.io/projected/deb01f64-2d7e-4d36-b660-8195a07ccf14-kube-api-access-xqm4m\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.206872 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb01f64-2d7e-4d36-b660-8195a07ccf14-serving-cert\") pod \"route-controller-manager-7855b9575f-8wtzn\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.346602 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.743467 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.794395 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/684ccdc3-1bcb-4227-8679-b9d31a21198c-kubelet-dir\") pod \"684ccdc3-1bcb-4227-8679-b9d31a21198c\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.794524 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/684ccdc3-1bcb-4227-8679-b9d31a21198c-kube-api-access\") pod \"684ccdc3-1bcb-4227-8679-b9d31a21198c\" (UID: \"684ccdc3-1bcb-4227-8679-b9d31a21198c\") " Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.794542 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/684ccdc3-1bcb-4227-8679-b9d31a21198c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "684ccdc3-1bcb-4227-8679-b9d31a21198c" (UID: "684ccdc3-1bcb-4227-8679-b9d31a21198c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.794906 4843 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/684ccdc3-1bcb-4227-8679-b9d31a21198c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.797818 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn"] Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.804740 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/684ccdc3-1bcb-4227-8679-b9d31a21198c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "684ccdc3-1bcb-4227-8679-b9d31a21198c" (UID: "684ccdc3-1bcb-4227-8679-b9d31a21198c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:20 crc kubenswrapper[4843]: W0314 09:13:20.813993 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeb01f64_2d7e_4d36_b660_8195a07ccf14.slice/crio-0570b282d818331f7f5206124e478e8c61953e7b0bf4c071ef70c14364699026 WatchSource:0}: Error finding container 0570b282d818331f7f5206124e478e8c61953e7b0bf4c071ef70c14364699026: Status 404 returned error can't find the container with id 0570b282d818331f7f5206124e478e8c61953e7b0bf4c071ef70c14364699026 Mar 14 09:13:20 crc kubenswrapper[4843]: I0314 09:13:20.895829 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/684ccdc3-1bcb-4227-8679-b9d31a21198c-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.451835 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"684ccdc3-1bcb-4227-8679-b9d31a21198c","Type":"ContainerDied","Data":"35386648194bcd292c97ec691bd21d7e1a30dee90f9631cd3e678ac55493f178"} Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.452651 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35386648194bcd292c97ec691bd21d7e1a30dee90f9631cd3e678ac55493f178" Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.452509 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.454855 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" event={"ID":"deb01f64-2d7e-4d36-b660-8195a07ccf14","Type":"ContainerStarted","Data":"8007471be3985ce04c3cc982dd0b9a32d9a73c24f543f66a1d2a865d8684c73a"} Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.454895 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" event={"ID":"deb01f64-2d7e-4d36-b660-8195a07ccf14","Type":"ContainerStarted","Data":"0570b282d818331f7f5206124e478e8c61953e7b0bf4c071ef70c14364699026"} Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.455314 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.461489 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.471068 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" podStartSLOduration=6.47104826 podStartE2EDuration="6.47104826s" podCreationTimestamp="2026-03-14 09:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:21.469756204 +0000 UTC m=+168.782367352" watchObservedRunningTime="2026-03-14 09:13:21.47104826 +0000 UTC m=+168.783659388" Mar 14 09:13:21 crc kubenswrapper[4843]: I0314 09:13:21.837920 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jj2ff"] Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.224686 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 14 09:13:24 crc kubenswrapper[4843]: E0314 09:13:24.224943 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="684ccdc3-1bcb-4227-8679-b9d31a21198c" containerName="pruner" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.224956 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="684ccdc3-1bcb-4227-8679-b9d31a21198c" containerName="pruner" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.225053 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="684ccdc3-1bcb-4227-8679-b9d31a21198c" containerName="pruner" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.225496 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.227309 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.228934 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.234644 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-var-lock\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.234717 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6032032-cdfd-41cd-8783-04efb653d09f-kube-api-access\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.234747 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.235493 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.336488 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.336616 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-var-lock\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.336673 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6032032-cdfd-41cd-8783-04efb653d09f-kube-api-access\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.336667 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.336703 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-var-lock\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.362418 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6032032-cdfd-41cd-8783-04efb653d09f-kube-api-access\") pod \"installer-9-crc\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.549096 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:13:24 crc kubenswrapper[4843]: I0314 09:13:24.972624 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 14 09:13:25 crc kubenswrapper[4843]: I0314 09:13:25.479736 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d6032032-cdfd-41cd-8783-04efb653d09f","Type":"ContainerStarted","Data":"91800f67d7419329edcb7b312788c43be13290e8d4db8f35fe633fd95cbca96f"} Mar 14 09:13:25 crc kubenswrapper[4843]: I0314 09:13:25.480189 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d6032032-cdfd-41cd-8783-04efb653d09f","Type":"ContainerStarted","Data":"a156136d3b737cf4007f79dd35cd940b9c98965f54487acc2d83dc8878d218ed"} Mar 14 09:13:25 crc kubenswrapper[4843]: I0314 09:13:25.498206 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.498182256 podStartE2EDuration="1.498182256s" podCreationTimestamp="2026-03-14 09:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:25.495937534 +0000 UTC m=+172.808548682" watchObservedRunningTime="2026-03-14 09:13:25.498182256 +0000 UTC m=+172.810793384" Mar 14 09:13:29 crc kubenswrapper[4843]: I0314 09:13:29.502507 4843 generic.go:334] "Generic (PLEG): container finished" podID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerID="92370561e6ec03fda689fe0ed7da1c04a650f66083a411f53f9c4b074c614602" exitCode=0 Mar 14 09:13:29 crc kubenswrapper[4843]: I0314 09:13:29.502599 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzj4r" event={"ID":"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f","Type":"ContainerDied","Data":"92370561e6ec03fda689fe0ed7da1c04a650f66083a411f53f9c4b074c614602"} Mar 14 09:13:29 crc kubenswrapper[4843]: I0314 09:13:29.506476 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerStarted","Data":"d3c353b93ad2313b0aba3dfecf63cffd97ddcfaa89c037bbba31f6cb088da2be"} Mar 14 09:13:29 crc kubenswrapper[4843]: I0314 09:13:29.510197 4843 generic.go:334] "Generic (PLEG): container finished" podID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerID="5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8" exitCode=0 Mar 14 09:13:29 crc kubenswrapper[4843]: I0314 09:13:29.510232 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx9cb" event={"ID":"3b7c5681-b6b0-4bfd-97d8-741096bb89d0","Type":"ContainerDied","Data":"5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8"} Mar 14 09:13:30 crc kubenswrapper[4843]: I0314 09:13:30.517724 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzj4r" event={"ID":"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f","Type":"ContainerStarted","Data":"2bf4eb7984dd563980a00e89f566f8f2e609c773ddfe82f9489a601032fa90e9"} Mar 14 09:13:30 crc kubenswrapper[4843]: I0314 09:13:30.519386 4843 generic.go:334] "Generic (PLEG): container finished" podID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerID="d3c353b93ad2313b0aba3dfecf63cffd97ddcfaa89c037bbba31f6cb088da2be" exitCode=0 Mar 14 09:13:30 crc kubenswrapper[4843]: I0314 09:13:30.519423 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerDied","Data":"d3c353b93ad2313b0aba3dfecf63cffd97ddcfaa89c037bbba31f6cb088da2be"} Mar 14 09:13:30 crc kubenswrapper[4843]: I0314 09:13:30.521923 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerStarted","Data":"64caeac5ab5881c7e1ca7063ed3696ae459e48f02896bbc9d8d559522fd3ab67"} Mar 14 09:13:30 crc kubenswrapper[4843]: I0314 09:13:30.526792 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx9cb" event={"ID":"3b7c5681-b6b0-4bfd-97d8-741096bb89d0","Type":"ContainerStarted","Data":"907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2"} Mar 14 09:13:30 crc kubenswrapper[4843]: I0314 09:13:30.541221 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kzj4r" podStartSLOduration=2.12973928 podStartE2EDuration="53.541197148s" podCreationTimestamp="2026-03-14 09:12:37 +0000 UTC" firstStartedPulling="2026-03-14 09:12:38.829237583 +0000 UTC m=+126.141848711" lastFinishedPulling="2026-03-14 09:13:30.240695451 +0000 UTC m=+177.553306579" observedRunningTime="2026-03-14 09:13:30.538588435 +0000 UTC m=+177.851199563" watchObservedRunningTime="2026-03-14 09:13:30.541197148 +0000 UTC m=+177.853808276" Mar 14 09:13:31 crc kubenswrapper[4843]: I0314 09:13:31.375226 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zx9cb" podStartSLOduration=3.031066853 podStartE2EDuration="54.3751985s" podCreationTimestamp="2026-03-14 09:12:37 +0000 UTC" firstStartedPulling="2026-03-14 09:12:38.807872179 +0000 UTC m=+126.120483317" lastFinishedPulling="2026-03-14 09:13:30.152003836 +0000 UTC m=+177.464614964" observedRunningTime="2026-03-14 09:13:30.58016533 +0000 UTC m=+177.892776448" watchObservedRunningTime="2026-03-14 09:13:31.3751985 +0000 UTC m=+178.687809628" Mar 14 09:13:31 crc kubenswrapper[4843]: I0314 09:13:31.534580 4843 generic.go:334] "Generic (PLEG): container finished" podID="f36123e7-f666-4382-8324-252e94a27884" containerID="64caeac5ab5881c7e1ca7063ed3696ae459e48f02896bbc9d8d559522fd3ab67" exitCode=0 Mar 14 09:13:31 crc kubenswrapper[4843]: I0314 09:13:31.534622 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerDied","Data":"64caeac5ab5881c7e1ca7063ed3696ae459e48f02896bbc9d8d559522fd3ab67"} Mar 14 09:13:35 crc kubenswrapper[4843]: I0314 09:13:35.708969 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b6856db49-hcfzr"] Mar 14 09:13:35 crc kubenswrapper[4843]: I0314 09:13:35.710106 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" podUID="4157111b-6a9f-4047-83c8-02c6de822519" containerName="controller-manager" containerID="cri-o://5694aba519048bc1f73ff8c6745b47b7465f7617160983d62cf8b8f0f1c71d74" gracePeriod=30 Mar 14 09:13:35 crc kubenswrapper[4843]: I0314 09:13:35.743341 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn"] Mar 14 09:13:35 crc kubenswrapper[4843]: I0314 09:13:35.743608 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" podUID="deb01f64-2d7e-4d36-b660-8195a07ccf14" containerName="route-controller-manager" containerID="cri-o://8007471be3985ce04c3cc982dd0b9a32d9a73c24f543f66a1d2a865d8684c73a" gracePeriod=30 Mar 14 09:13:36 crc kubenswrapper[4843]: I0314 09:13:36.567017 4843 generic.go:334] "Generic (PLEG): container finished" podID="4157111b-6a9f-4047-83c8-02c6de822519" containerID="5694aba519048bc1f73ff8c6745b47b7465f7617160983d62cf8b8f0f1c71d74" exitCode=0 Mar 14 09:13:36 crc kubenswrapper[4843]: I0314 09:13:36.567134 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" event={"ID":"4157111b-6a9f-4047-83c8-02c6de822519","Type":"ContainerDied","Data":"5694aba519048bc1f73ff8c6745b47b7465f7617160983d62cf8b8f0f1c71d74"} Mar 14 09:13:36 crc kubenswrapper[4843]: I0314 09:13:36.569012 4843 generic.go:334] "Generic (PLEG): container finished" podID="deb01f64-2d7e-4d36-b660-8195a07ccf14" containerID="8007471be3985ce04c3cc982dd0b9a32d9a73c24f543f66a1d2a865d8684c73a" exitCode=0 Mar 14 09:13:36 crc kubenswrapper[4843]: I0314 09:13:36.569054 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" event={"ID":"deb01f64-2d7e-4d36-b660-8195a07ccf14","Type":"ContainerDied","Data":"8007471be3985ce04c3cc982dd0b9a32d9a73c24f543f66a1d2a865d8684c73a"} Mar 14 09:13:37 crc kubenswrapper[4843]: I0314 09:13:37.870917 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:13:37 crc kubenswrapper[4843]: I0314 09:13:37.871257 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.069206 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.069246 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.224898 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.229016 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.256916 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2"] Mar 14 09:13:38 crc kubenswrapper[4843]: E0314 09:13:38.257156 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb01f64-2d7e-4d36-b660-8195a07ccf14" containerName="route-controller-manager" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.257168 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb01f64-2d7e-4d36-b660-8195a07ccf14" containerName="route-controller-manager" Mar 14 09:13:38 crc kubenswrapper[4843]: E0314 09:13:38.257185 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4157111b-6a9f-4047-83c8-02c6de822519" containerName="controller-manager" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.257192 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4157111b-6a9f-4047-83c8-02c6de822519" containerName="controller-manager" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.257315 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4157111b-6a9f-4047-83c8-02c6de822519" containerName="controller-manager" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.257329 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb01f64-2d7e-4d36-b660-8195a07ccf14" containerName="route-controller-manager" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.257766 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.269856 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2"] Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.322536 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-client-ca\") pod \"deb01f64-2d7e-4d36-b660-8195a07ccf14\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.322841 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-config\") pod \"deb01f64-2d7e-4d36-b660-8195a07ccf14\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.322950 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sfwr\" (UniqueName: \"kubernetes.io/projected/4157111b-6a9f-4047-83c8-02c6de822519-kube-api-access-6sfwr\") pod \"4157111b-6a9f-4047-83c8-02c6de822519\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323057 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqm4m\" (UniqueName: \"kubernetes.io/projected/deb01f64-2d7e-4d36-b660-8195a07ccf14-kube-api-access-xqm4m\") pod \"deb01f64-2d7e-4d36-b660-8195a07ccf14\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323146 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-client-ca\") pod \"4157111b-6a9f-4047-83c8-02c6de822519\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323572 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-config\") pod \"4157111b-6a9f-4047-83c8-02c6de822519\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323661 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb01f64-2d7e-4d36-b660-8195a07ccf14-serving-cert\") pod \"deb01f64-2d7e-4d36-b660-8195a07ccf14\" (UID: \"deb01f64-2d7e-4d36-b660-8195a07ccf14\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323753 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-proxy-ca-bundles\") pod \"4157111b-6a9f-4047-83c8-02c6de822519\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323297 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-client-ca" (OuterVolumeSpecName: "client-ca") pod "deb01f64-2d7e-4d36-b660-8195a07ccf14" (UID: "deb01f64-2d7e-4d36-b660-8195a07ccf14"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323403 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-config" (OuterVolumeSpecName: "config") pod "deb01f64-2d7e-4d36-b660-8195a07ccf14" (UID: "deb01f64-2d7e-4d36-b660-8195a07ccf14"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.323906 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4157111b-6a9f-4047-83c8-02c6de822519-serving-cert\") pod \"4157111b-6a9f-4047-83c8-02c6de822519\" (UID: \"4157111b-6a9f-4047-83c8-02c6de822519\") " Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.324129 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-client-ca" (OuterVolumeSpecName: "client-ca") pod "4157111b-6a9f-4047-83c8-02c6de822519" (UID: "4157111b-6a9f-4047-83c8-02c6de822519"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.324158 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.324203 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-config" (OuterVolumeSpecName: "config") pod "4157111b-6a9f-4047-83c8-02c6de822519" (UID: "4157111b-6a9f-4047-83c8-02c6de822519"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.324223 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb01f64-2d7e-4d36-b660-8195a07ccf14-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.324280 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4157111b-6a9f-4047-83c8-02c6de822519" (UID: "4157111b-6a9f-4047-83c8-02c6de822519"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.329354 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4157111b-6a9f-4047-83c8-02c6de822519-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4157111b-6a9f-4047-83c8-02c6de822519" (UID: "4157111b-6a9f-4047-83c8-02c6de822519"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.330178 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb01f64-2d7e-4d36-b660-8195a07ccf14-kube-api-access-xqm4m" (OuterVolumeSpecName: "kube-api-access-xqm4m") pod "deb01f64-2d7e-4d36-b660-8195a07ccf14" (UID: "deb01f64-2d7e-4d36-b660-8195a07ccf14"). InnerVolumeSpecName "kube-api-access-xqm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.331400 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb01f64-2d7e-4d36-b660-8195a07ccf14-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "deb01f64-2d7e-4d36-b660-8195a07ccf14" (UID: "deb01f64-2d7e-4d36-b660-8195a07ccf14"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.339374 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4157111b-6a9f-4047-83c8-02c6de822519-kube-api-access-6sfwr" (OuterVolumeSpecName: "kube-api-access-6sfwr") pod "4157111b-6a9f-4047-83c8-02c6de822519" (UID: "4157111b-6a9f-4047-83c8-02c6de822519"). InnerVolumeSpecName "kube-api-access-6sfwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.425799 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-config\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.425887 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-client-ca\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.425942 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8bc81d-1b65-4d5f-b54f-f674800b395b-serving-cert\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426050 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnp2k\" (UniqueName: \"kubernetes.io/projected/2a8bc81d-1b65-4d5f-b54f-f674800b395b-kube-api-access-mnp2k\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426264 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sfwr\" (UniqueName: \"kubernetes.io/projected/4157111b-6a9f-4047-83c8-02c6de822519-kube-api-access-6sfwr\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426327 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqm4m\" (UniqueName: \"kubernetes.io/projected/deb01f64-2d7e-4d36-b660-8195a07ccf14-kube-api-access-xqm4m\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426346 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426364 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426382 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb01f64-2d7e-4d36-b660-8195a07ccf14-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426400 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4157111b-6a9f-4047-83c8-02c6de822519-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.426418 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4157111b-6a9f-4047-83c8-02c6de822519-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.528251 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-config\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.528330 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-client-ca\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.528364 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8bc81d-1b65-4d5f-b54f-f674800b395b-serving-cert\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.528402 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnp2k\" (UniqueName: \"kubernetes.io/projected/2a8bc81d-1b65-4d5f-b54f-f674800b395b-kube-api-access-mnp2k\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.530882 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-config\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.531213 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-client-ca\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.538507 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8bc81d-1b65-4d5f-b54f-f674800b395b-serving-cert\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.547810 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnp2k\" (UniqueName: \"kubernetes.io/projected/2a8bc81d-1b65-4d5f-b54f-f674800b395b-kube-api-access-mnp2k\") pod \"route-controller-manager-bd8664ff8-rspn2\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.581524 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" event={"ID":"4157111b-6a9f-4047-83c8-02c6de822519","Type":"ContainerDied","Data":"b3b29f2d3ce60d61ac7760293902286f039a22e793aa409904c76b3de3202ab3"} Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.581640 4843 scope.go:117] "RemoveContainer" containerID="5694aba519048bc1f73ff8c6745b47b7465f7617160983d62cf8b8f0f1c71d74" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.581797 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.585048 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" event={"ID":"deb01f64-2d7e-4d36-b660-8195a07ccf14","Type":"ContainerDied","Data":"0570b282d818331f7f5206124e478e8c61953e7b0bf4c071ef70c14364699026"} Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.585202 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.590695 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.617114 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b6856db49-hcfzr"] Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.624086 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b6856db49-hcfzr"] Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.631940 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn"] Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.636772 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7855b9575f-8wtzn"] Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.770019 4843 patch_prober.go:28] interesting pod/controller-manager-6b6856db49-hcfzr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 14 09:13:38 crc kubenswrapper[4843]: I0314 09:13:38.770253 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6b6856db49-hcfzr" podUID="4157111b-6a9f-4047-83c8-02c6de822519" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.236573 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.236826 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.288040 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.295204 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.346923 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4157111b-6a9f-4047-83c8-02c6de822519" path="/var/lib/kubelet/pods/4157111b-6a9f-4047-83c8-02c6de822519/volumes" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.347548 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb01f64-2d7e-4d36-b660-8195a07ccf14" path="/var/lib/kubelet/pods/deb01f64-2d7e-4d36-b660-8195a07ccf14/volumes" Mar 14 09:13:39 crc kubenswrapper[4843]: I0314 09:13:39.857366 4843 scope.go:117] "RemoveContainer" containerID="8007471be3985ce04c3cc982dd0b9a32d9a73c24f543f66a1d2a865d8684c73a" Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.424797 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2"] Mar 14 09:13:40 crc kubenswrapper[4843]: W0314 09:13:40.461346 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a8bc81d_1b65_4d5f_b54f_f674800b395b.slice/crio-5b2c51b835658b61bf7a1a74933713239cb1d2b6bb57cae754336382a2756b4b WatchSource:0}: Error finding container 5b2c51b835658b61bf7a1a74933713239cb1d2b6bb57cae754336382a2756b4b: Status 404 returned error can't find the container with id 5b2c51b835658b61bf7a1a74933713239cb1d2b6bb57cae754336382a2756b4b Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.600637 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerID="b2b7a85bc0131426f55ffe652dd79762a2397f02cb671b3cecd668d347487dbe" exitCode=0 Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.600717 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mss9d" event={"ID":"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb","Type":"ContainerDied","Data":"b2b7a85bc0131426f55ffe652dd79762a2397f02cb671b3cecd668d347487dbe"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.604032 4843 generic.go:334] "Generic (PLEG): container finished" podID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerID="16ecf23b13ebe1948e58504e7c66c64ba731b5b28fc8c14551e7432f9e3b649a" exitCode=0 Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.604114 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xg5hq" event={"ID":"59b75c6f-f60e-4ed2-8dec-d61a29121a63","Type":"ContainerDied","Data":"16ecf23b13ebe1948e58504e7c66c64ba731b5b28fc8c14551e7432f9e3b649a"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.610245 4843 generic.go:334] "Generic (PLEG): container finished" podID="8220c353-b74c-4703-ac79-af96379e89fe" containerID="c407df7c6a78977956bb2212f97a460d5af3d7abb779e5d102cad33b1272e18f" exitCode=0 Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.610333 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmpq6" event={"ID":"8220c353-b74c-4703-ac79-af96379e89fe","Type":"ContainerDied","Data":"c407df7c6a78977956bb2212f97a460d5af3d7abb779e5d102cad33b1272e18f"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.614466 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerStarted","Data":"059126c0943c1deeeca3f9a01bcce95c7ad3745e4f30430be71e684ad6fd1048"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.621768 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerStarted","Data":"ccdb5a14b19869408b7b07b1062541044849e5659a0a0af4d5b5cf845cd0f0c5"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.625431 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" event={"ID":"2a8bc81d-1b65-4d5f-b54f-f674800b395b","Type":"ContainerStarted","Data":"5b2c51b835658b61bf7a1a74933713239cb1d2b6bb57cae754336382a2756b4b"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.632816 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerStarted","Data":"57cf099bf0adae4642557f0162a505ab0a88e5ce692b512a1a0fa6d5189eb8e8"} Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.645893 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vrzt2" podStartSLOduration=3.670655629 podStartE2EDuration="1m3.645867976s" podCreationTimestamp="2026-03-14 09:12:37 +0000 UTC" firstStartedPulling="2026-03-14 09:12:39.875599628 +0000 UTC m=+127.188210756" lastFinishedPulling="2026-03-14 09:13:39.850811965 +0000 UTC m=+187.163423103" observedRunningTime="2026-03-14 09:13:40.642884912 +0000 UTC m=+187.955496050" watchObservedRunningTime="2026-03-14 09:13:40.645867976 +0000 UTC m=+187.958479114" Mar 14 09:13:40 crc kubenswrapper[4843]: I0314 09:13:40.684200 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9lwgw" podStartSLOduration=5.698922096 podStartE2EDuration="59.684182458s" podCreationTimestamp="2026-03-14 09:12:41 +0000 UTC" firstStartedPulling="2026-03-14 09:12:42.962809933 +0000 UTC m=+130.275421061" lastFinishedPulling="2026-03-14 09:13:36.948070295 +0000 UTC m=+184.260681423" observedRunningTime="2026-03-14 09:13:40.683336025 +0000 UTC m=+187.995947163" watchObservedRunningTime="2026-03-14 09:13:40.684182458 +0000 UTC m=+187.996793586" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.038538 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-766fd86c54-t4jcr"] Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.039182 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.042112 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.042183 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.044343 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.045344 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.045496 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.046244 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.063079 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.076018 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-766fd86c54-t4jcr"] Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.167464 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-config\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.167502 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-client-ca\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.167522 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-proxy-ca-bundles\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.167551 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5l4v\" (UniqueName: \"kubernetes.io/projected/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-kube-api-access-d5l4v\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.167572 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-serving-cert\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.269054 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-config\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.269100 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-client-ca\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.269118 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-proxy-ca-bundles\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.269148 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5l4v\" (UniqueName: \"kubernetes.io/projected/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-kube-api-access-d5l4v\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.269169 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-serving-cert\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.272406 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-config\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.273079 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-client-ca\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.273379 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-proxy-ca-bundles\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.278129 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-serving-cert\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.289691 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5l4v\" (UniqueName: \"kubernetes.io/projected/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-kube-api-access-d5l4v\") pod \"controller-manager-766fd86c54-t4jcr\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.374600 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.499357 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.499409 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.584920 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zx9cb"] Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.585222 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zx9cb" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="registry-server" containerID="cri-o://907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2" gracePeriod=2 Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.605208 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-766fd86c54-t4jcr"] Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.641734 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mss9d" event={"ID":"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb","Type":"ContainerStarted","Data":"1d2523b4024629aa058e87c4903c05bb59ed654faa17d84400b74924e537fb45"} Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.646840 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xg5hq" event={"ID":"59b75c6f-f60e-4ed2-8dec-d61a29121a63","Type":"ContainerStarted","Data":"6f952c674a9fc176b0001131a1b35e3f122f4db6be4dbb98da66aea15a6819c9"} Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.647896 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" event={"ID":"2a8bc81d-1b65-4d5f-b54f-f674800b395b","Type":"ContainerStarted","Data":"d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a"} Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.648078 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.651178 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmpq6" event={"ID":"8220c353-b74c-4703-ac79-af96379e89fe","Type":"ContainerStarted","Data":"a32ae428a788540ef28f5251d0de6014324adbc9ee5b3deb8cf76358dfda7ab6"} Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.663208 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" event={"ID":"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1","Type":"ContainerStarted","Data":"646f2e04078f672a6ba4e921df12378f3480c7384850c9bf42d55034fcd33660"} Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.669824 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.676341 4843 generic.go:334] "Generic (PLEG): container finished" podID="3568e895-1a84-487a-9042-376d2bb9e28f" containerID="57cf099bf0adae4642557f0162a505ab0a88e5ce692b512a1a0fa6d5189eb8e8" exitCode=0 Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.677049 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerDied","Data":"57cf099bf0adae4642557f0162a505ab0a88e5ce692b512a1a0fa6d5189eb8e8"} Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.681086 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xg5hq" podStartSLOduration=4.641155552 podStartE2EDuration="1m2.681074103s" podCreationTimestamp="2026-03-14 09:12:39 +0000 UTC" firstStartedPulling="2026-03-14 09:12:42.989797897 +0000 UTC m=+130.302409015" lastFinishedPulling="2026-03-14 09:13:41.029716438 +0000 UTC m=+188.342327566" observedRunningTime="2026-03-14 09:13:41.680435015 +0000 UTC m=+188.993046143" watchObservedRunningTime="2026-03-14 09:13:41.681074103 +0000 UTC m=+188.993685231" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.682557 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mss9d" podStartSLOduration=2.471012144 podStartE2EDuration="1m4.682551054s" podCreationTimestamp="2026-03-14 09:12:37 +0000 UTC" firstStartedPulling="2026-03-14 09:12:38.841228494 +0000 UTC m=+126.153839612" lastFinishedPulling="2026-03-14 09:13:41.052767384 +0000 UTC m=+188.365378522" observedRunningTime="2026-03-14 09:13:41.663361797 +0000 UTC m=+188.975972925" watchObservedRunningTime="2026-03-14 09:13:41.682551054 +0000 UTC m=+188.995162182" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.706252 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" podStartSLOduration=6.706228898 podStartE2EDuration="6.706228898s" podCreationTimestamp="2026-03-14 09:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:41.705101316 +0000 UTC m=+189.017712444" watchObservedRunningTime="2026-03-14 09:13:41.706228898 +0000 UTC m=+189.018840026" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.752871 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mmpq6" podStartSLOduration=2.6626559419999998 podStartE2EDuration="1m2.752851654s" podCreationTimestamp="2026-03-14 09:12:39 +0000 UTC" firstStartedPulling="2026-03-14 09:12:40.92147873 +0000 UTC m=+128.234089858" lastFinishedPulling="2026-03-14 09:13:41.011674432 +0000 UTC m=+188.324285570" observedRunningTime="2026-03-14 09:13:41.729656024 +0000 UTC m=+189.042267142" watchObservedRunningTime="2026-03-14 09:13:41.752851654 +0000 UTC m=+189.065462782" Mar 14 09:13:41 crc kubenswrapper[4843]: I0314 09:13:41.962372 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.088890 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-utilities\") pod \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.088948 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lhj4\" (UniqueName: \"kubernetes.io/projected/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-kube-api-access-7lhj4\") pod \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.089105 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-catalog-content\") pod \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\" (UID: \"3b7c5681-b6b0-4bfd-97d8-741096bb89d0\") " Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.098600 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-utilities" (OuterVolumeSpecName: "utilities") pod "3b7c5681-b6b0-4bfd-97d8-741096bb89d0" (UID: "3b7c5681-b6b0-4bfd-97d8-741096bb89d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.103983 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-kube-api-access-7lhj4" (OuterVolumeSpecName: "kube-api-access-7lhj4") pod "3b7c5681-b6b0-4bfd-97d8-741096bb89d0" (UID: "3b7c5681-b6b0-4bfd-97d8-741096bb89d0"). InnerVolumeSpecName "kube-api-access-7lhj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.166388 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b7c5681-b6b0-4bfd-97d8-741096bb89d0" (UID: "3b7c5681-b6b0-4bfd-97d8-741096bb89d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.190696 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.190733 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.190743 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lhj4\" (UniqueName: \"kubernetes.io/projected/3b7c5681-b6b0-4bfd-97d8-741096bb89d0-kube-api-access-7lhj4\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.552745 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9lwgw" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="registry-server" probeResult="failure" output=< Mar 14 09:13:42 crc kubenswrapper[4843]: timeout: failed to connect service ":50051" within 1s Mar 14 09:13:42 crc kubenswrapper[4843]: > Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.682669 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" event={"ID":"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1","Type":"ContainerStarted","Data":"edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48"} Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.682877 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.685570 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerStarted","Data":"25bf24975d63a934c63c123325ef1ca1d58ee2581e50f26a6ffce3fecf7b9cb3"} Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.687640 4843 generic.go:334] "Generic (PLEG): container finished" podID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerID="907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2" exitCode=0 Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.687889 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zx9cb" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.688013 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx9cb" event={"ID":"3b7c5681-b6b0-4bfd-97d8-741096bb89d0","Type":"ContainerDied","Data":"907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2"} Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.688041 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zx9cb" event={"ID":"3b7c5681-b6b0-4bfd-97d8-741096bb89d0","Type":"ContainerDied","Data":"02ca0dd7fb363d2a992b7a208943e194d21266d94339fdfec2781ee4931410b8"} Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.688062 4843 scope.go:117] "RemoveContainer" containerID="907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.696393 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.703379 4843 scope.go:117] "RemoveContainer" containerID="5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.708380 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" podStartSLOduration=7.7083644289999995 podStartE2EDuration="7.708364429s" podCreationTimestamp="2026-03-14 09:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:42.705442717 +0000 UTC m=+190.018053845" watchObservedRunningTime="2026-03-14 09:13:42.708364429 +0000 UTC m=+190.020975557" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.730609 4843 scope.go:117] "RemoveContainer" containerID="12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.745818 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zx9cb"] Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.747133 4843 scope.go:117] "RemoveContainer" containerID="907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2" Mar 14 09:13:42 crc kubenswrapper[4843]: E0314 09:13:42.747724 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2\": container with ID starting with 907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2 not found: ID does not exist" containerID="907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.747764 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2"} err="failed to get container status \"907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2\": rpc error: code = NotFound desc = could not find container \"907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2\": container with ID starting with 907e64192fa22985a720a3dc89e8a56a16ef6107038460f9561cfe973bb45db2 not found: ID does not exist" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.747793 4843 scope.go:117] "RemoveContainer" containerID="5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8" Mar 14 09:13:42 crc kubenswrapper[4843]: E0314 09:13:42.748105 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8\": container with ID starting with 5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8 not found: ID does not exist" containerID="5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.748131 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8"} err="failed to get container status \"5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8\": rpc error: code = NotFound desc = could not find container \"5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8\": container with ID starting with 5e0dbe3465bf101bde4aeb5f2963cb6ba16c564be0889ba468662b9340addfc8 not found: ID does not exist" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.748151 4843 scope.go:117] "RemoveContainer" containerID="12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45" Mar 14 09:13:42 crc kubenswrapper[4843]: E0314 09:13:42.748351 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45\": container with ID starting with 12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45 not found: ID does not exist" containerID="12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.748374 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45"} err="failed to get container status \"12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45\": rpc error: code = NotFound desc = could not find container \"12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45\": container with ID starting with 12f930ed58c8937d6c5498e3de0eb283d92f358b42d9bd7ac7583c83d0c5fc45 not found: ID does not exist" Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.753855 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zx9cb"] Mar 14 09:13:42 crc kubenswrapper[4843]: I0314 09:13:42.789121 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jfzjg" podStartSLOduration=3.637757949 podStartE2EDuration="1m2.789087011s" podCreationTimestamp="2026-03-14 09:12:40 +0000 UTC" firstStartedPulling="2026-03-14 09:12:42.952786231 +0000 UTC m=+130.265397359" lastFinishedPulling="2026-03-14 09:13:42.104115303 +0000 UTC m=+189.416726421" observedRunningTime="2026-03-14 09:13:42.773564706 +0000 UTC m=+190.086175834" watchObservedRunningTime="2026-03-14 09:13:42.789087011 +0000 UTC m=+190.101698209" Mar 14 09:13:43 crc kubenswrapper[4843]: I0314 09:13:43.346905 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" path="/var/lib/kubelet/pods/3b7c5681-b6b0-4bfd-97d8-741096bb89d0/volumes" Mar 14 09:13:46 crc kubenswrapper[4843]: I0314 09:13:46.869719 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" podUID="664b59b6-607d-4320-8802-0ce25a001245" containerName="oauth-openshift" containerID="cri-o://a25d079c2109343dc441f5552d4a894d4c543f2c9165b1c10ad955e4261aa8a3" gracePeriod=15 Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.450369 4843 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod4c222d32-5588-40c0-a714-dc50cd379a67"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod4c222d32-5588-40c0-a714-dc50cd379a67] : Timed out while waiting for systemd to remove kubepods-burstable-pod4c222d32_5588_40c0_a714_dc50cd379a67.slice" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.663830 4843 ???:1] "http: TLS handshake error from 192.168.126.11:40510: no serving certificate available for the kubelet" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.687332 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.687401 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.716815 4843 generic.go:334] "Generic (PLEG): container finished" podID="664b59b6-607d-4320-8802-0ce25a001245" containerID="a25d079c2109343dc441f5552d4a894d4c543f2c9165b1c10ad955e4261aa8a3" exitCode=0 Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.716855 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" event={"ID":"664b59b6-607d-4320-8802-0ce25a001245","Type":"ContainerDied","Data":"a25d079c2109343dc441f5552d4a894d4c543f2c9165b1c10ad955e4261aa8a3"} Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.733081 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.773991 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.816588 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965378 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-error\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965445 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cnzn\" (UniqueName: \"kubernetes.io/projected/664b59b6-607d-4320-8802-0ce25a001245-kube-api-access-5cnzn\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965497 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-router-certs\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965529 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-trusted-ca-bundle\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965562 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-provider-selection\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965599 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/664b59b6-607d-4320-8802-0ce25a001245-audit-dir\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965625 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965650 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-cliconfig\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965671 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-session\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.965792 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/664b59b6-607d-4320-8802-0ce25a001245-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966447 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966585 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966633 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966795 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-login\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966837 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-idp-0-file-data\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966867 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-serving-cert\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966920 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-audit-policies\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.966958 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-ocp-branding-template\") pod \"664b59b6-607d-4320-8802-0ce25a001245\" (UID: \"664b59b6-607d-4320-8802-0ce25a001245\") " Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.967263 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.967315 4843 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/664b59b6-607d-4320-8802-0ce25a001245-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.967328 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.967343 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.969767 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.973044 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.976128 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.976295 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/664b59b6-607d-4320-8802-0ce25a001245-kube-api-access-5cnzn" (OuterVolumeSpecName: "kube-api-access-5cnzn") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "kube-api-access-5cnzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.976408 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.976645 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.978654 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.981799 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.986432 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:47 crc kubenswrapper[4843]: I0314 09:13:47.987659 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "664b59b6-607d-4320-8802-0ce25a001245" (UID: "664b59b6-607d-4320-8802-0ce25a001245"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069138 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069185 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069198 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069212 4843 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/664b59b6-607d-4320-8802-0ce25a001245-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069228 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069242 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069254 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cnzn\" (UniqueName: \"kubernetes.io/projected/664b59b6-607d-4320-8802-0ce25a001245-kube-api-access-5cnzn\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069266 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069305 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.069319 4843 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/664b59b6-607d-4320-8802-0ce25a001245-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.298556 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.298642 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.349238 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.725117 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" event={"ID":"664b59b6-607d-4320-8802-0ce25a001245","Type":"ContainerDied","Data":"39f35b2172e35d29da84d499355a75d44302b47bc74c4b8623de2d76dbeaff5d"} Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.725195 4843 scope.go:117] "RemoveContainer" containerID="a25d079c2109343dc441f5552d4a894d4c543f2c9165b1c10ad955e4261aa8a3" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.725527 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jj2ff" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.799186 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.805060 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jj2ff"] Mar 14 09:13:48 crc kubenswrapper[4843]: I0314 09:13:48.812343 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jj2ff"] Mar 14 09:13:49 crc kubenswrapper[4843]: I0314 09:13:49.352614 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="664b59b6-607d-4320-8802-0ce25a001245" path="/var/lib/kubelet/pods/664b59b6-607d-4320-8802-0ce25a001245/volumes" Mar 14 09:13:49 crc kubenswrapper[4843]: I0314 09:13:49.686683 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:13:49 crc kubenswrapper[4843]: I0314 09:13:49.686759 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:13:49 crc kubenswrapper[4843]: I0314 09:13:49.725175 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:13:49 crc kubenswrapper[4843]: I0314 09:13:49.776399 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:13:50 crc kubenswrapper[4843]: I0314 09:13:50.082133 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:13:50 crc kubenswrapper[4843]: I0314 09:13:50.082214 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:13:50 crc kubenswrapper[4843]: I0314 09:13:50.131545 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:13:50 crc kubenswrapper[4843]: I0314 09:13:50.587595 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrzt2"] Mar 14 09:13:50 crc kubenswrapper[4843]: I0314 09:13:50.739753 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vrzt2" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="registry-server" containerID="cri-o://ccdb5a14b19869408b7b07b1062541044849e5659a0a0af4d5b5cf845cd0f0c5" gracePeriod=2 Mar 14 09:13:50 crc kubenswrapper[4843]: I0314 09:13:50.827568 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:13:51 crc kubenswrapper[4843]: I0314 09:13:51.107993 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:13:51 crc kubenswrapper[4843]: I0314 09:13:51.108083 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:13:51 crc kubenswrapper[4843]: I0314 09:13:51.181667 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:13:51 crc kubenswrapper[4843]: I0314 09:13:51.564760 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:13:51 crc kubenswrapper[4843]: I0314 09:13:51.609339 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:13:51 crc kubenswrapper[4843]: I0314 09:13:51.800565 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:13:52 crc kubenswrapper[4843]: I0314 09:13:52.758581 4843 generic.go:334] "Generic (PLEG): container finished" podID="f36123e7-f666-4382-8324-252e94a27884" containerID="ccdb5a14b19869408b7b07b1062541044849e5659a0a0af4d5b5cf845cd0f0c5" exitCode=0 Mar 14 09:13:52 crc kubenswrapper[4843]: I0314 09:13:52.759388 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerDied","Data":"ccdb5a14b19869408b7b07b1062541044849e5659a0a0af4d5b5cf845cd0f0c5"} Mar 14 09:13:52 crc kubenswrapper[4843]: I0314 09:13:52.916108 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:13:52 crc kubenswrapper[4843]: I0314 09:13:52.977678 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xg5hq"] Mar 14 09:13:52 crc kubenswrapper[4843]: I0314 09:13:52.977873 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xg5hq" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="registry-server" containerID="cri-o://6f952c674a9fc176b0001131a1b35e3f122f4db6be4dbb98da66aea15a6819c9" gracePeriod=2 Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.044742 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj"] Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.044946 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="registry-server" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.044957 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="registry-server" Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.044965 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="registry-server" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.044970 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="registry-server" Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.044986 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="extract-content" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.044994 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="extract-content" Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.045002 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="extract-content" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045009 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="extract-content" Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.045021 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="extract-utilities" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045027 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="extract-utilities" Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.045037 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="extract-utilities" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045043 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="extract-utilities" Mar 14 09:13:53 crc kubenswrapper[4843]: E0314 09:13:53.045051 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="664b59b6-607d-4320-8802-0ce25a001245" containerName="oauth-openshift" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045057 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="664b59b6-607d-4320-8802-0ce25a001245" containerName="oauth-openshift" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045141 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b7c5681-b6b0-4bfd-97d8-741096bb89d0" containerName="registry-server" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045152 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="664b59b6-607d-4320-8802-0ce25a001245" containerName="oauth-openshift" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045163 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36123e7-f666-4382-8324-252e94a27884" containerName="registry-server" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.045520 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.048391 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.048956 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.049784 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.050503 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.050769 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.051910 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-utilities\") pod \"f36123e7-f666-4382-8324-252e94a27884\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052580 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krq99\" (UniqueName: \"kubernetes.io/projected/f36123e7-f666-4382-8324-252e94a27884-kube-api-access-krq99\") pod \"f36123e7-f666-4382-8324-252e94a27884\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052623 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-catalog-content\") pod \"f36123e7-f666-4382-8324-252e94a27884\" (UID: \"f36123e7-f666-4382-8324-252e94a27884\") " Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052646 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-utilities" (OuterVolumeSpecName: "utilities") pod "f36123e7-f666-4382-8324-252e94a27884" (UID: "f36123e7-f666-4382-8324-252e94a27884"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052764 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052785 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052808 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-error\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052826 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052843 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44986e29-0f22-4f42-a902-2f7772310d2b-audit-dir\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052860 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-session\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052889 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052923 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj7m8\" (UniqueName: \"kubernetes.io/projected/44986e29-0f22-4f42-a902-2f7772310d2b-kube-api-access-jj7m8\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052943 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.052982 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.053016 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-audit-policies\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.053035 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.053052 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-login\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.053083 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.053119 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.055365 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.055524 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.055615 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.055658 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.055674 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.055774 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.071145 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f36123e7-f666-4382-8324-252e94a27884-kube-api-access-krq99" (OuterVolumeSpecName: "kube-api-access-krq99") pod "f36123e7-f666-4382-8324-252e94a27884" (UID: "f36123e7-f666-4382-8324-252e94a27884"). InnerVolumeSpecName "kube-api-access-krq99". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.093701 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.101184 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj"] Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.101352 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.106067 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.109249 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.124909 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f36123e7-f666-4382-8324-252e94a27884" (UID: "f36123e7-f666-4382-8324-252e94a27884"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154572 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154650 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-audit-policies\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154673 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154715 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-login\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154743 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154798 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154872 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154896 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-error\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154934 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154957 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44986e29-0f22-4f42-a902-2f7772310d2b-audit-dir\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.154976 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-session\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155055 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155108 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj7m8\" (UniqueName: \"kubernetes.io/projected/44986e29-0f22-4f42-a902-2f7772310d2b-kube-api-access-jj7m8\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155129 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155184 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krq99\" (UniqueName: \"kubernetes.io/projected/f36123e7-f666-4382-8324-252e94a27884-kube-api-access-krq99\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155197 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36123e7-f666-4382-8324-252e94a27884-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155620 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44986e29-0f22-4f42-a902-2f7772310d2b-audit-dir\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.155736 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-audit-policies\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.156321 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.156725 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.159127 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-error\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.159156 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.160230 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.160772 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.160940 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.161547 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-template-login\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.161652 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.162704 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.165805 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/44986e29-0f22-4f42-a902-2f7772310d2b-v4-0-config-system-session\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.175749 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj7m8\" (UniqueName: \"kubernetes.io/projected/44986e29-0f22-4f42-a902-2f7772310d2b-kube-api-access-jj7m8\") pod \"oauth-openshift-5f94f6fdf4-n49wj\" (UID: \"44986e29-0f22-4f42-a902-2f7772310d2b\") " pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.377428 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.765649 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrzt2" event={"ID":"f36123e7-f666-4382-8324-252e94a27884","Type":"ContainerDied","Data":"3e6567d361d83eb85b5c195629f4136eef79fde8b6b4fbbebc8bb7c975bfaf9f"} Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.766118 4843 scope.go:117] "RemoveContainer" containerID="ccdb5a14b19869408b7b07b1062541044849e5659a0a0af4d5b5cf845cd0f0c5" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.765806 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrzt2" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.784700 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrzt2"] Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.785560 4843 scope.go:117] "RemoveContainer" containerID="64caeac5ab5881c7e1ca7063ed3696ae459e48f02896bbc9d8d559522fd3ab67" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.789634 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vrzt2"] Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.835310 4843 scope.go:117] "RemoveContainer" containerID="39fbee9e3534a3ba4a6ba3d977073f5515ed8b7c35c85a9aab284b1d5ffd87a3" Mar 14 09:13:53 crc kubenswrapper[4843]: I0314 09:13:53.860919 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj"] Mar 14 09:13:53 crc kubenswrapper[4843]: W0314 09:13:53.928325 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44986e29_0f22_4f42_a902_2f7772310d2b.slice/crio-20c721130a6d500db527075cee2237a04864f3168c82f41dd502b9410b494538 WatchSource:0}: Error finding container 20c721130a6d500db527075cee2237a04864f3168c82f41dd502b9410b494538: Status 404 returned error can't find the container with id 20c721130a6d500db527075cee2237a04864f3168c82f41dd502b9410b494538 Mar 14 09:13:54 crc kubenswrapper[4843]: I0314 09:13:54.778561 4843 generic.go:334] "Generic (PLEG): container finished" podID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerID="6f952c674a9fc176b0001131a1b35e3f122f4db6be4dbb98da66aea15a6819c9" exitCode=0 Mar 14 09:13:54 crc kubenswrapper[4843]: I0314 09:13:54.778608 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xg5hq" event={"ID":"59b75c6f-f60e-4ed2-8dec-d61a29121a63","Type":"ContainerDied","Data":"6f952c674a9fc176b0001131a1b35e3f122f4db6be4dbb98da66aea15a6819c9"} Mar 14 09:13:54 crc kubenswrapper[4843]: I0314 09:13:54.785830 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" event={"ID":"44986e29-0f22-4f42-a902-2f7772310d2b","Type":"ContainerStarted","Data":"20c721130a6d500db527075cee2237a04864f3168c82f41dd502b9410b494538"} Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.169777 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.288685 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-catalog-content\") pod \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.288782 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-utilities\") pod \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.288855 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbvwl\" (UniqueName: \"kubernetes.io/projected/59b75c6f-f60e-4ed2-8dec-d61a29121a63-kube-api-access-hbvwl\") pod \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\" (UID: \"59b75c6f-f60e-4ed2-8dec-d61a29121a63\") " Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.290976 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-utilities" (OuterVolumeSpecName: "utilities") pod "59b75c6f-f60e-4ed2-8dec-d61a29121a63" (UID: "59b75c6f-f60e-4ed2-8dec-d61a29121a63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.299527 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b75c6f-f60e-4ed2-8dec-d61a29121a63-kube-api-access-hbvwl" (OuterVolumeSpecName: "kube-api-access-hbvwl") pod "59b75c6f-f60e-4ed2-8dec-d61a29121a63" (UID: "59b75c6f-f60e-4ed2-8dec-d61a29121a63"). InnerVolumeSpecName "kube-api-access-hbvwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.312264 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59b75c6f-f60e-4ed2-8dec-d61a29121a63" (UID: "59b75c6f-f60e-4ed2-8dec-d61a29121a63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.346468 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f36123e7-f666-4382-8324-252e94a27884" path="/var/lib/kubelet/pods/f36123e7-f666-4382-8324-252e94a27884/volumes" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.379829 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9lwgw"] Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.380058 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9lwgw" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="registry-server" containerID="cri-o://059126c0943c1deeeca3f9a01bcce95c7ad3745e4f30430be71e684ad6fd1048" gracePeriod=2 Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.390039 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.390080 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbvwl\" (UniqueName: \"kubernetes.io/projected/59b75c6f-f60e-4ed2-8dec-d61a29121a63-kube-api-access-hbvwl\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.390093 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b75c6f-f60e-4ed2-8dec-d61a29121a63-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.714147 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-766fd86c54-t4jcr"] Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.715052 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" podUID="7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" containerName="controller-manager" containerID="cri-o://edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48" gracePeriod=30 Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.733799 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2"] Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.734032 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" podUID="2a8bc81d-1b65-4d5f-b54f-f674800b395b" containerName="route-controller-manager" containerID="cri-o://d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a" gracePeriod=30 Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.794823 4843 generic.go:334] "Generic (PLEG): container finished" podID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerID="059126c0943c1deeeca3f9a01bcce95c7ad3745e4f30430be71e684ad6fd1048" exitCode=0 Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.794901 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerDied","Data":"059126c0943c1deeeca3f9a01bcce95c7ad3745e4f30430be71e684ad6fd1048"} Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.796911 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xg5hq" event={"ID":"59b75c6f-f60e-4ed2-8dec-d61a29121a63","Type":"ContainerDied","Data":"92ba406da930481e2a77149d259e1eac1d46876f8c7f706f28095abc37ae93df"} Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.796949 4843 scope.go:117] "RemoveContainer" containerID="6f952c674a9fc176b0001131a1b35e3f122f4db6be4dbb98da66aea15a6819c9" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.797086 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xg5hq" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.798456 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" event={"ID":"44986e29-0f22-4f42-a902-2f7772310d2b","Type":"ContainerStarted","Data":"9b0252c84debfae04f94172901e273f9fd170c30dda7171e8d239d35049e8f05"} Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.798730 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.888168 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.908907 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5f94f6fdf4-n49wj" podStartSLOduration=34.908867759 podStartE2EDuration="34.908867759s" podCreationTimestamp="2026-03-14 09:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:55.819765841 +0000 UTC m=+203.132376969" watchObservedRunningTime="2026-03-14 09:13:55.908867759 +0000 UTC m=+203.221478877" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.925146 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.930287 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xg5hq"] Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.935344 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xg5hq"] Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.939712 4843 scope.go:117] "RemoveContainer" containerID="16ecf23b13ebe1948e58504e7c66c64ba731b5b28fc8c14551e7432f9e3b649a" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.961025 4843 scope.go:117] "RemoveContainer" containerID="faf12fe2b9e8ce822d8b8317c6c5779ec87701f1200c3723ec65e3ac3f79ad88" Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.997126 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-utilities\") pod \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.997191 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsscf\" (UniqueName: \"kubernetes.io/projected/92538af0-1d79-4c19-b7c2-f1c1b89595a3-kube-api-access-tsscf\") pod \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.997249 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-catalog-content\") pod \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\" (UID: \"92538af0-1d79-4c19-b7c2-f1c1b89595a3\") " Mar 14 09:13:55 crc kubenswrapper[4843]: I0314 09:13:55.998354 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-utilities" (OuterVolumeSpecName: "utilities") pod "92538af0-1d79-4c19-b7c2-f1c1b89595a3" (UID: "92538af0-1d79-4c19-b7c2-f1c1b89595a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.006449 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92538af0-1d79-4c19-b7c2-f1c1b89595a3-kube-api-access-tsscf" (OuterVolumeSpecName: "kube-api-access-tsscf") pod "92538af0-1d79-4c19-b7c2-f1c1b89595a3" (UID: "92538af0-1d79-4c19-b7c2-f1c1b89595a3"). InnerVolumeSpecName "kube-api-access-tsscf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.098214 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsscf\" (UniqueName: \"kubernetes.io/projected/92538af0-1d79-4c19-b7c2-f1c1b89595a3-kube-api-access-tsscf\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.098244 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.181133 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92538af0-1d79-4c19-b7c2-f1c1b89595a3" (UID: "92538af0-1d79-4c19-b7c2-f1c1b89595a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.198874 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92538af0-1d79-4c19-b7c2-f1c1b89595a3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.294252 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.341943 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.401479 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8bc81d-1b65-4d5f-b54f-f674800b395b-serving-cert\") pod \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.401543 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnp2k\" (UniqueName: \"kubernetes.io/projected/2a8bc81d-1b65-4d5f-b54f-f674800b395b-kube-api-access-mnp2k\") pod \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.401624 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-config\") pod \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.401702 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-client-ca\") pod \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\" (UID: \"2a8bc81d-1b65-4d5f-b54f-f674800b395b\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.402627 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-client-ca" (OuterVolumeSpecName: "client-ca") pod "2a8bc81d-1b65-4d5f-b54f-f674800b395b" (UID: "2a8bc81d-1b65-4d5f-b54f-f674800b395b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.402842 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-config" (OuterVolumeSpecName: "config") pod "2a8bc81d-1b65-4d5f-b54f-f674800b395b" (UID: "2a8bc81d-1b65-4d5f-b54f-f674800b395b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.406902 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a8bc81d-1b65-4d5f-b54f-f674800b395b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2a8bc81d-1b65-4d5f-b54f-f674800b395b" (UID: "2a8bc81d-1b65-4d5f-b54f-f674800b395b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.406945 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a8bc81d-1b65-4d5f-b54f-f674800b395b-kube-api-access-mnp2k" (OuterVolumeSpecName: "kube-api-access-mnp2k") pod "2a8bc81d-1b65-4d5f-b54f-f674800b395b" (UID: "2a8bc81d-1b65-4d5f-b54f-f674800b395b"). InnerVolumeSpecName "kube-api-access-mnp2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.502670 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5l4v\" (UniqueName: \"kubernetes.io/projected/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-kube-api-access-d5l4v\") pod \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.502749 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-proxy-ca-bundles\") pod \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.502807 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-config\") pod \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.502851 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-client-ca\") pod \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.502903 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-serving-cert\") pod \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\" (UID: \"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1\") " Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.503180 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.503198 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8bc81d-1b65-4d5f-b54f-f674800b395b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.503212 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnp2k\" (UniqueName: \"kubernetes.io/projected/2a8bc81d-1b65-4d5f-b54f-f674800b395b-kube-api-access-mnp2k\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.503226 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8bc81d-1b65-4d5f-b54f-f674800b395b-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.504371 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-config" (OuterVolumeSpecName: "config") pod "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" (UID: "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.504469 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-client-ca" (OuterVolumeSpecName: "client-ca") pod "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" (UID: "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.504496 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" (UID: "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.506185 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" (UID: "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.506238 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-kube-api-access-d5l4v" (OuterVolumeSpecName: "kube-api-access-d5l4v") pod "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" (UID: "7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1"). InnerVolumeSpecName "kube-api-access-d5l4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.604685 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.604739 4843 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.604757 4843 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.604774 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5l4v\" (UniqueName: \"kubernetes.io/projected/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-kube-api-access-d5l4v\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.604791 4843 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.808777 4843 generic.go:334] "Generic (PLEG): container finished" podID="2a8bc81d-1b65-4d5f-b54f-f674800b395b" containerID="d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a" exitCode=0 Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.808826 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.808825 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" event={"ID":"2a8bc81d-1b65-4d5f-b54f-f674800b395b","Type":"ContainerDied","Data":"d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a"} Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.808948 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2" event={"ID":"2a8bc81d-1b65-4d5f-b54f-f674800b395b","Type":"ContainerDied","Data":"5b2c51b835658b61bf7a1a74933713239cb1d2b6bb57cae754336382a2756b4b"} Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.808974 4843 scope.go:117] "RemoveContainer" containerID="d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.811503 4843 generic.go:334] "Generic (PLEG): container finished" podID="7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" containerID="edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48" exitCode=0 Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.811553 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.811582 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" event={"ID":"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1","Type":"ContainerDied","Data":"edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48"} Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.811661 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-766fd86c54-t4jcr" event={"ID":"7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1","Type":"ContainerDied","Data":"646f2e04078f672a6ba4e921df12378f3480c7384850c9bf42d55034fcd33660"} Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.815401 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9lwgw" event={"ID":"92538af0-1d79-4c19-b7c2-f1c1b89595a3","Type":"ContainerDied","Data":"fc481e852e9ff72029931ba9c71dadf2f199c6ac34e743ab01c566463da2ec43"} Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.815479 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9lwgw" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.830721 4843 scope.go:117] "RemoveContainer" containerID="d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a" Mar 14 09:13:56 crc kubenswrapper[4843]: E0314 09:13:56.831430 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a\": container with ID starting with d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a not found: ID does not exist" containerID="d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.831492 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a"} err="failed to get container status \"d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a\": rpc error: code = NotFound desc = could not find container \"d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a\": container with ID starting with d6024cb87bbc026a3c9147999b8a35f0340810c8c7297c43fda71d480fd65e5a not found: ID does not exist" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.831646 4843 scope.go:117] "RemoveContainer" containerID="edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.836933 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2"] Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.841379 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bd8664ff8-rspn2"] Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.851506 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-766fd86c54-t4jcr"] Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.866471 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-766fd86c54-t4jcr"] Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.869410 4843 scope.go:117] "RemoveContainer" containerID="edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48" Mar 14 09:13:56 crc kubenswrapper[4843]: E0314 09:13:56.870145 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48\": container with ID starting with edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48 not found: ID does not exist" containerID="edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.870193 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48"} err="failed to get container status \"edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48\": rpc error: code = NotFound desc = could not find container \"edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48\": container with ID starting with edb44ad45648485cc7cea7440c6ee9440ed12304f2783081a08294999426ef48 not found: ID does not exist" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.870219 4843 scope.go:117] "RemoveContainer" containerID="059126c0943c1deeeca3f9a01bcce95c7ad3745e4f30430be71e684ad6fd1048" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.872801 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9lwgw"] Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.875431 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9lwgw"] Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.888852 4843 scope.go:117] "RemoveContainer" containerID="d3c353b93ad2313b0aba3dfecf63cffd97ddcfaa89c037bbba31f6cb088da2be" Mar 14 09:13:56 crc kubenswrapper[4843]: I0314 09:13:56.907307 4843 scope.go:117] "RemoveContainer" containerID="6a78abfcee61c1388010cca35e259c2321f39887f19380ae91695dcc2822023d" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.054824 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9"] Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055165 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" containerName="controller-manager" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055181 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" containerName="controller-manager" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055200 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="extract-utilities" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055207 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="extract-utilities" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055222 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="extract-content" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055229 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="extract-content" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055240 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="extract-utilities" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055247 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="extract-utilities" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055257 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="extract-content" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055265 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="extract-content" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055295 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="registry-server" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055302 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="registry-server" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055311 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8bc81d-1b65-4d5f-b54f-f674800b395b" containerName="route-controller-manager" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055320 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8bc81d-1b65-4d5f-b54f-f674800b395b" containerName="route-controller-manager" Mar 14 09:13:57 crc kubenswrapper[4843]: E0314 09:13:57.055329 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="registry-server" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055338 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="registry-server" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055448 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" containerName="registry-server" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055457 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a8bc81d-1b65-4d5f-b54f-f674800b395b" containerName="route-controller-manager" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055465 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" containerName="controller-manager" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.055510 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" containerName="registry-server" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.056173 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.057676 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6fbf776b56-8ntpp"] Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.058726 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.060081 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.062346 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.065602 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.065963 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.066217 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.066615 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.066720 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.066945 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.072644 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.072804 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.073114 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.073264 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.081861 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9"] Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.082595 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.087895 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fbf776b56-8ntpp"] Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.211801 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-proxy-ca-bundles\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212154 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-config\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212319 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzj7p\" (UniqueName: \"kubernetes.io/projected/ab867447-e512-4976-b8fe-c7d23ace44cf-kube-api-access-fzj7p\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212443 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-config\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212551 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-serving-cert\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212647 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-client-ca\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212802 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr5th\" (UniqueName: \"kubernetes.io/projected/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-kube-api-access-mr5th\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.212887 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab867447-e512-4976-b8fe-c7d23ace44cf-serving-cert\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.213007 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-client-ca\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314439 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-client-ca\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314536 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-proxy-ca-bundles\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314609 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-config\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314666 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzj7p\" (UniqueName: \"kubernetes.io/projected/ab867447-e512-4976-b8fe-c7d23ace44cf-kube-api-access-fzj7p\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314744 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-config\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314807 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-serving-cert\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.314868 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-client-ca\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.315003 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr5th\" (UniqueName: \"kubernetes.io/projected/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-kube-api-access-mr5th\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.315055 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab867447-e512-4976-b8fe-c7d23ace44cf-serving-cert\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.316546 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-client-ca\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.317727 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-client-ca\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.318501 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-config\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.319155 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab867447-e512-4976-b8fe-c7d23ace44cf-proxy-ca-bundles\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.319793 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-config\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.322257 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-serving-cert\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.322911 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab867447-e512-4976-b8fe-c7d23ace44cf-serving-cert\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.332014 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr5th\" (UniqueName: \"kubernetes.io/projected/b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae-kube-api-access-mr5th\") pod \"route-controller-manager-59954f54c6-bnvq9\" (UID: \"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae\") " pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.341503 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzj7p\" (UniqueName: \"kubernetes.io/projected/ab867447-e512-4976-b8fe-c7d23ace44cf-kube-api-access-fzj7p\") pod \"controller-manager-6fbf776b56-8ntpp\" (UID: \"ab867447-e512-4976-b8fe-c7d23ace44cf\") " pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.348180 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a8bc81d-1b65-4d5f-b54f-f674800b395b" path="/var/lib/kubelet/pods/2a8bc81d-1b65-4d5f-b54f-f674800b395b/volumes" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.348816 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b75c6f-f60e-4ed2-8dec-d61a29121a63" path="/var/lib/kubelet/pods/59b75c6f-f60e-4ed2-8dec-d61a29121a63/volumes" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.349620 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1" path="/var/lib/kubelet/pods/7cc991fd-fe0b-4bbe-bb20-4eb0c78c81a1/volumes" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.350853 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92538af0-1d79-4c19-b7c2-f1c1b89595a3" path="/var/lib/kubelet/pods/92538af0-1d79-4c19-b7c2-f1c1b89595a3/volumes" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.435885 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.471629 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.762548 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fbf776b56-8ntpp"] Mar 14 09:13:57 crc kubenswrapper[4843]: W0314 09:13:57.767602 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab867447_e512_4976_b8fe_c7d23ace44cf.slice/crio-df4a6c105a995ded969e3ce71a0383e0dd9fd215accd2a9a431734e5cd0f3e22 WatchSource:0}: Error finding container df4a6c105a995ded969e3ce71a0383e0dd9fd215accd2a9a431734e5cd0f3e22: Status 404 returned error can't find the container with id df4a6c105a995ded969e3ce71a0383e0dd9fd215accd2a9a431734e5cd0f3e22 Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.832463 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" event={"ID":"ab867447-e512-4976-b8fe-c7d23ace44cf","Type":"ContainerStarted","Data":"df4a6c105a995ded969e3ce71a0383e0dd9fd215accd2a9a431734e5cd0f3e22"} Mar 14 09:13:57 crc kubenswrapper[4843]: I0314 09:13:57.905527 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9"] Mar 14 09:13:57 crc kubenswrapper[4843]: W0314 09:13:57.916306 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb25c6fc4_54d8_4412_95b1_cf5f7e2f01ae.slice/crio-6998c31afee2f12ee5063e94b8516f324f5c03f50e3d9dcf004e4e7527ca5a50 WatchSource:0}: Error finding container 6998c31afee2f12ee5063e94b8516f324f5c03f50e3d9dcf004e4e7527ca5a50: Status 404 returned error can't find the container with id 6998c31afee2f12ee5063e94b8516f324f5c03f50e3d9dcf004e4e7527ca5a50 Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.838929 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" event={"ID":"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae","Type":"ContainerStarted","Data":"878218ee3b94672b39b458b2153b60168ce31572c73245e30c21fd79b54f1938"} Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.838973 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" event={"ID":"b25c6fc4-54d8-4412-95b1-cf5f7e2f01ae","Type":"ContainerStarted","Data":"6998c31afee2f12ee5063e94b8516f324f5c03f50e3d9dcf004e4e7527ca5a50"} Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.839421 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.841620 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" event={"ID":"ab867447-e512-4976-b8fe-c7d23ace44cf","Type":"ContainerStarted","Data":"3b17a3b07ddece567cfcc915651f4b245b04362f955a54bb502be367394cc5f9"} Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.841867 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.846704 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.852508 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.868159 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59954f54c6-bnvq9" podStartSLOduration=3.868138454 podStartE2EDuration="3.868138454s" podCreationTimestamp="2026-03-14 09:13:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:58.86416307 +0000 UTC m=+206.176774208" watchObservedRunningTime="2026-03-14 09:13:58.868138454 +0000 UTC m=+206.180749582" Mar 14 09:13:58 crc kubenswrapper[4843]: I0314 09:13:58.910672 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6fbf776b56-8ntpp" podStartSLOduration=3.910647935 podStartE2EDuration="3.910647935s" podCreationTimestamp="2026-03-14 09:13:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:13:58.904690673 +0000 UTC m=+206.217301801" watchObservedRunningTime="2026-03-14 09:13:58.910647935 +0000 UTC m=+206.223259063" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.134941 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557994-q425x"] Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.137350 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.140350 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.140674 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.140815 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.153887 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557994-q425x"] Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.256323 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxjlp\" (UniqueName: \"kubernetes.io/projected/f190b327-cd1f-4a28-9d0b-7e1980a334a6-kube-api-access-rxjlp\") pod \"auto-csr-approver-29557994-q425x\" (UID: \"f190b327-cd1f-4a28-9d0b-7e1980a334a6\") " pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.358250 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxjlp\" (UniqueName: \"kubernetes.io/projected/f190b327-cd1f-4a28-9d0b-7e1980a334a6-kube-api-access-rxjlp\") pod \"auto-csr-approver-29557994-q425x\" (UID: \"f190b327-cd1f-4a28-9d0b-7e1980a334a6\") " pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.378941 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxjlp\" (UniqueName: \"kubernetes.io/projected/f190b327-cd1f-4a28-9d0b-7e1980a334a6-kube-api-access-rxjlp\") pod \"auto-csr-approver-29557994-q425x\" (UID: \"f190b327-cd1f-4a28-9d0b-7e1980a334a6\") " pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.462717 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:00 crc kubenswrapper[4843]: I0314 09:14:00.896717 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557994-q425x"] Mar 14 09:14:00 crc kubenswrapper[4843]: W0314 09:14:00.907489 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf190b327_cd1f_4a28_9d0b_7e1980a334a6.slice/crio-37d38bd47abcaed0d47efe58435b51503799b5705e24c9e00d5185ca31db19a0 WatchSource:0}: Error finding container 37d38bd47abcaed0d47efe58435b51503799b5705e24c9e00d5185ca31db19a0: Status 404 returned error can't find the container with id 37d38bd47abcaed0d47efe58435b51503799b5705e24c9e00d5185ca31db19a0 Mar 14 09:14:01 crc kubenswrapper[4843]: I0314 09:14:01.861941 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557994-q425x" event={"ID":"f190b327-cd1f-4a28-9d0b-7e1980a334a6","Type":"ContainerStarted","Data":"37d38bd47abcaed0d47efe58435b51503799b5705e24c9e00d5185ca31db19a0"} Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.908081 4843 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.908554 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://19b57e62b0b1cbb6a7798bc999bfe302198c071e3447232baebdb5e2d8af69e6" gracePeriod=15 Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.908651 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://da2bed11209848aa1fce423d8809128a64e548ac1395de019d1ca614d4973423" gracePeriod=15 Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.908770 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f29b51f4b235799e889cec4242cf9895f2ac9b154a1e4e1a4eaaea4ae9d4366e" gracePeriod=15 Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.908818 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8a4aa9fcdb00dbc4dcb3869d76cbfdfede25f06447edefc02107913ab84faf02" gracePeriod=15 Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.908995 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://9fd4021be66e4d56e2f5bfc26c24aacc52eaa4eb7c38fdda57c52422f7d9fd68" gracePeriod=15 Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.910820 4843 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.912976 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913032 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913100 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913113 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913132 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913144 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913162 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913175 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913220 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913232 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913247 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913258 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913292 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913304 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913322 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913334 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.913352 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913363 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913604 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913629 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913644 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913657 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913679 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913699 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913712 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.913726 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.914087 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.914109 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.914367 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.917933 4843 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.919457 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.924903 4843 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 14 09:14:02 crc kubenswrapper[4843]: E0314 09:14:02.993401 4843 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.995874 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.995980 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.996029 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.996059 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.996104 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.996386 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.996475 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:02 crc kubenswrapper[4843]: I0314 09:14:02.996546 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.028317 4843 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.029250 4843 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.029775 4843 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.030254 4843 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.030621 4843 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.030668 4843 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.030993 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="200ms" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098361 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098435 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098458 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098474 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098496 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098533 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098548 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098547 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098625 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098572 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098670 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098698 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098728 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098724 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098774 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.098866 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.232376 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="400ms" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.294675 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.341888 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.398508 4843 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" volumeName="registry-storage" Mar 14 09:14:03 crc kubenswrapper[4843]: E0314 09:14:03.633520 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="800ms" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.877903 4843 generic.go:334] "Generic (PLEG): container finished" podID="d6032032-cdfd-41cd-8783-04efb653d09f" containerID="91800f67d7419329edcb7b312788c43be13290e8d4db8f35fe633fd95cbca96f" exitCode=0 Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.877974 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d6032032-cdfd-41cd-8783-04efb653d09f","Type":"ContainerDied","Data":"91800f67d7419329edcb7b312788c43be13290e8d4db8f35fe633fd95cbca96f"} Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.878946 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.879612 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.883465 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.886609 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.888242 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f29b51f4b235799e889cec4242cf9895f2ac9b154a1e4e1a4eaaea4ae9d4366e" exitCode=0 Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.888288 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8a4aa9fcdb00dbc4dcb3869d76cbfdfede25f06447edefc02107913ab84faf02" exitCode=0 Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.888306 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da2bed11209848aa1fce423d8809128a64e548ac1395de019d1ca614d4973423" exitCode=0 Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.888317 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9fd4021be66e4d56e2f5bfc26c24aacc52eaa4eb7c38fdda57c52422f7d9fd68" exitCode=2 Mar 14 09:14:03 crc kubenswrapper[4843]: I0314 09:14:03.888392 4843 scope.go:117] "RemoveContainer" containerID="5d681cc31691c3d1bc8a43e5b5defff5af05fc3d69005078c809a816c11484ec" Mar 14 09:14:04 crc kubenswrapper[4843]: E0314 09:14:04.435116 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="1.6s" Mar 14 09:14:05 crc kubenswrapper[4843]: E0314 09:14:05.166568 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-19b57e62b0b1cbb6a7798bc999bfe302198c071e3447232baebdb5e2d8af69e6.scope\": RecentStats: unable to find data in memory cache]" Mar 14 09:14:05 crc kubenswrapper[4843]: I0314 09:14:05.905056 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 14 09:14:05 crc kubenswrapper[4843]: I0314 09:14:05.905889 4843 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="19b57e62b0b1cbb6a7798bc999bfe302198c071e3447232baebdb5e2d8af69e6" exitCode=0 Mar 14 09:14:06 crc kubenswrapper[4843]: E0314 09:14:06.036070 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="3.2s" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.914259 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d6032032-cdfd-41cd-8783-04efb653d09f","Type":"ContainerDied","Data":"a156136d3b737cf4007f79dd35cd940b9c98965f54487acc2d83dc8878d218ed"} Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.914326 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a156136d3b737cf4007f79dd35cd940b9c98965f54487acc2d83dc8878d218ed" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.951352 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.952586 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.953100 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.958240 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.959186 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.959807 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.960183 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:06 crc kubenswrapper[4843]: I0314 09:14:06.961043 4843 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:06 crc kubenswrapper[4843]: W0314 09:14:06.972466 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-355d3e8b71d181af301f35a95c05484e765edf56d0f82cfdd4e9ae06afaf1d56 WatchSource:0}: Error finding container 355d3e8b71d181af301f35a95c05484e765edf56d0f82cfdd4e9ae06afaf1d56: Status 404 returned error can't find the container with id 355d3e8b71d181af301f35a95c05484e765edf56d0f82cfdd4e9ae06afaf1d56 Mar 14 09:14:06 crc kubenswrapper[4843]: E0314 09:14:06.989656 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189caa57b72da910 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:14:06.9892324 +0000 UTC m=+214.301843548,LastTimestamp:2026-03-14 09:14:06.9892324 +0000 UTC m=+214.301843548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.144849 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.144989 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145499 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-kubelet-dir\") pod \"d6032032-cdfd-41cd-8783-04efb653d09f\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145559 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d6032032-cdfd-41cd-8783-04efb653d09f" (UID: "d6032032-cdfd-41cd-8783-04efb653d09f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145599 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145658 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-var-lock\") pod \"d6032032-cdfd-41cd-8783-04efb653d09f\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145683 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145712 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6032032-cdfd-41cd-8783-04efb653d09f-kube-api-access\") pod \"d6032032-cdfd-41cd-8783-04efb653d09f\" (UID: \"d6032032-cdfd-41cd-8783-04efb653d09f\") " Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145679 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145690 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-var-lock" (OuterVolumeSpecName: "var-lock") pod "d6032032-cdfd-41cd-8783-04efb653d09f" (UID: "d6032032-cdfd-41cd-8783-04efb653d09f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145700 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145908 4843 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145925 4843 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145938 4843 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d6032032-cdfd-41cd-8783-04efb653d09f-var-lock\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145948 4843 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.145958 4843 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.151218 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6032032-cdfd-41cd-8783-04efb653d09f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d6032032-cdfd-41cd-8783-04efb653d09f" (UID: "d6032032-cdfd-41cd-8783-04efb653d09f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.246738 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6032032-cdfd-41cd-8783-04efb653d09f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.346841 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.921773 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"84e0d02f7ef73b46bd1c83fa8e8b48f8e8fc29051d90a3b3edd8ea1d71a76130"} Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.921821 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"355d3e8b71d181af301f35a95c05484e765edf56d0f82cfdd4e9ae06afaf1d56"} Mar 14 09:14:07 crc kubenswrapper[4843]: E0314 09:14:07.922541 4843 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.922899 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.923210 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.924574 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.925432 4843 scope.go:117] "RemoveContainer" containerID="f29b51f4b235799e889cec4242cf9895f2ac9b154a1e4e1a4eaaea4ae9d4366e" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.925482 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.926067 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.926409 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.926664 4843 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.927767 4843 generic.go:334] "Generic (PLEG): container finished" podID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" containerID="5bca3d22e3db02690b7de5f6b625153e094fafb851e9b5c4f149e659213683b8" exitCode=0 Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.927818 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557994-q425x" event={"ID":"f190b327-cd1f-4a28-9d0b-7e1980a334a6","Type":"ContainerDied","Data":"5bca3d22e3db02690b7de5f6b625153e094fafb851e9b5c4f149e659213683b8"} Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.927855 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.928493 4843 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.928828 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.929147 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.944068 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.945009 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.945474 4843 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.946482 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.946756 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.947004 4843 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.959572 4843 scope.go:117] "RemoveContainer" containerID="8a4aa9fcdb00dbc4dcb3869d76cbfdfede25f06447edefc02107913ab84faf02" Mar 14 09:14:07 crc kubenswrapper[4843]: I0314 09:14:07.979223 4843 scope.go:117] "RemoveContainer" containerID="da2bed11209848aa1fce423d8809128a64e548ac1395de019d1ca614d4973423" Mar 14 09:14:08 crc kubenswrapper[4843]: I0314 09:14:08.000418 4843 scope.go:117] "RemoveContainer" containerID="9fd4021be66e4d56e2f5bfc26c24aacc52eaa4eb7c38fdda57c52422f7d9fd68" Mar 14 09:14:08 crc kubenswrapper[4843]: I0314 09:14:08.017164 4843 scope.go:117] "RemoveContainer" containerID="19b57e62b0b1cbb6a7798bc999bfe302198c071e3447232baebdb5e2d8af69e6" Mar 14 09:14:08 crc kubenswrapper[4843]: I0314 09:14:08.031761 4843 scope.go:117] "RemoveContainer" containerID="025e1eececd2799bfda2d8956afccae562ed75627c225a98db4a483e76433fbd" Mar 14 09:14:08 crc kubenswrapper[4843]: E0314 09:14:08.876494 4843 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189caa57b72da910 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-14 09:14:06.9892324 +0000 UTC m=+214.301843548,LastTimestamp:2026-03-14 09:14:06.9892324 +0000 UTC m=+214.301843548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 14 09:14:08 crc kubenswrapper[4843]: E0314 09:14:08.935763 4843 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:14:09 crc kubenswrapper[4843]: E0314 09:14:09.237592 4843 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="6.4s" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.273355 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.274091 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.274687 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.274992 4843 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.380415 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxjlp\" (UniqueName: \"kubernetes.io/projected/f190b327-cd1f-4a28-9d0b-7e1980a334a6-kube-api-access-rxjlp\") pod \"f190b327-cd1f-4a28-9d0b-7e1980a334a6\" (UID: \"f190b327-cd1f-4a28-9d0b-7e1980a334a6\") " Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.389191 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f190b327-cd1f-4a28-9d0b-7e1980a334a6-kube-api-access-rxjlp" (OuterVolumeSpecName: "kube-api-access-rxjlp") pod "f190b327-cd1f-4a28-9d0b-7e1980a334a6" (UID: "f190b327-cd1f-4a28-9d0b-7e1980a334a6"). InnerVolumeSpecName "kube-api-access-rxjlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.482504 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxjlp\" (UniqueName: \"kubernetes.io/projected/f190b327-cd1f-4a28-9d0b-7e1980a334a6-kube-api-access-rxjlp\") on node \"crc\" DevicePath \"\"" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.942136 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557994-q425x" event={"ID":"f190b327-cd1f-4a28-9d0b-7e1980a334a6","Type":"ContainerDied","Data":"37d38bd47abcaed0d47efe58435b51503799b5705e24c9e00d5185ca31db19a0"} Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.942226 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37d38bd47abcaed0d47efe58435b51503799b5705e24c9e00d5185ca31db19a0" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.942237 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557994-q425x" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.969230 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:09 crc kubenswrapper[4843]: I0314 09:14:09.969696 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: E0314 09:14:13.008487 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:14:13Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:14:13Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:14:13Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-14T09:14:13Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: E0314 09:14:13.010253 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: E0314 09:14:13.010772 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: E0314 09:14:13.011234 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: E0314 09:14:13.011582 4843 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: E0314 09:14:13.011634 4843 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 14 09:14:13 crc kubenswrapper[4843]: I0314 09:14:13.344240 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:13 crc kubenswrapper[4843]: I0314 09:14:13.344518 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.338484 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.340638 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.341197 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.357733 4843 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.357778 4843 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:14 crc kubenswrapper[4843]: E0314 09:14:14.358250 4843 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.359179 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:14 crc kubenswrapper[4843]: W0314 09:14:14.385586 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-ef64e3cc28dad139a3a0434de4496d2d08a843daf78c93649cf239e59bb30888 WatchSource:0}: Error finding container ef64e3cc28dad139a3a0434de4496d2d08a843daf78c93649cf239e59bb30888: Status 404 returned error can't find the container with id ef64e3cc28dad139a3a0434de4496d2d08a843daf78c93649cf239e59bb30888 Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.974511 4843 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="af7325e67ba44c15dc95debc86f3b4e90b9398adcb2e237789f30e6aa90c797f" exitCode=0 Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.974576 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"af7325e67ba44c15dc95debc86f3b4e90b9398adcb2e237789f30e6aa90c797f"} Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.974623 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef64e3cc28dad139a3a0434de4496d2d08a843daf78c93649cf239e59bb30888"} Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.975060 4843 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.975096 4843 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.976044 4843 status_manager.go:851] "Failed to get status for pod" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" pod="openshift-infra/auto-csr-approver-29557994-q425x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29557994-q425x\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:14 crc kubenswrapper[4843]: E0314 09:14:14.976045 4843 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:14 crc kubenswrapper[4843]: I0314 09:14:14.976637 4843 status_manager.go:851] "Failed to get status for pod" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.162:6443: connect: connection refused" Mar 14 09:14:15 crc kubenswrapper[4843]: I0314 09:14:15.387258 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:14:15 crc kubenswrapper[4843]: I0314 09:14:15.387741 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:14:15 crc kubenswrapper[4843]: I0314 09:14:15.989691 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"25176ae98938d6d662f700b3256097864f90c1af5a31b13aa96d2bea95a1e19f"} Mar 14 09:14:15 crc kubenswrapper[4843]: I0314 09:14:15.990146 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ea1ff5ddb874ffebf1cdaf537de750f4dc6a0daf700680a8ef094e9f817ff73c"} Mar 14 09:14:15 crc kubenswrapper[4843]: I0314 09:14:15.990160 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"079ed1930434966e84376f314390d0d9e7bea3841257ba4b860e47c5c76e8777"} Mar 14 09:14:16 crc kubenswrapper[4843]: I0314 09:14:16.996787 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 14 09:14:16 crc kubenswrapper[4843]: I0314 09:14:16.997670 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 14 09:14:16 crc kubenswrapper[4843]: I0314 09:14:16.997705 4843 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c2c83d7114544865d6fe66909ce703a28d8428bde328863ec8071bbb2e53adec" exitCode=1 Mar 14 09:14:16 crc kubenswrapper[4843]: I0314 09:14:16.997755 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c2c83d7114544865d6fe66909ce703a28d8428bde328863ec8071bbb2e53adec"} Mar 14 09:14:16 crc kubenswrapper[4843]: I0314 09:14:16.998159 4843 scope.go:117] "RemoveContainer" containerID="c2c83d7114544865d6fe66909ce703a28d8428bde328863ec8071bbb2e53adec" Mar 14 09:14:17 crc kubenswrapper[4843]: I0314 09:14:17.001028 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4e489b66b83a05e1b7627d003efc7579ba71a00a07b4fc9df276084bad3109d2"} Mar 14 09:14:17 crc kubenswrapper[4843]: I0314 09:14:17.001074 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fb4b43030c99bdea30aca7dbf63da3a73552664e7abe3e9877a8c4bf8d257f05"} Mar 14 09:14:17 crc kubenswrapper[4843]: I0314 09:14:17.001240 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:17 crc kubenswrapper[4843]: I0314 09:14:17.001372 4843 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:17 crc kubenswrapper[4843]: I0314 09:14:17.001401 4843 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:17 crc kubenswrapper[4843]: I0314 09:14:17.976513 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:14:18 crc kubenswrapper[4843]: I0314 09:14:18.009775 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 14 09:14:18 crc kubenswrapper[4843]: I0314 09:14:18.011225 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 14 09:14:18 crc kubenswrapper[4843]: I0314 09:14:18.011303 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b091dcfbbc7a1b6b0cb9f7493f48cde024d7f68280c8af7c1c6df38c61c8617c"} Mar 14 09:14:19 crc kubenswrapper[4843]: I0314 09:14:19.359906 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:19 crc kubenswrapper[4843]: I0314 09:14:19.359957 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:19 crc kubenswrapper[4843]: I0314 09:14:19.367892 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:19 crc kubenswrapper[4843]: I0314 09:14:19.775770 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:14:22 crc kubenswrapper[4843]: I0314 09:14:22.009814 4843 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:22 crc kubenswrapper[4843]: I0314 09:14:22.034822 4843 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:22 crc kubenswrapper[4843]: I0314 09:14:22.034850 4843 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:22 crc kubenswrapper[4843]: I0314 09:14:22.038801 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:23 crc kubenswrapper[4843]: I0314 09:14:23.042732 4843 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:23 crc kubenswrapper[4843]: I0314 09:14:23.042776 4843 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="476cb746-a5a7-4db7-97b6-0f4936e2aded" Mar 14 09:14:23 crc kubenswrapper[4843]: I0314 09:14:23.361206 4843 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="95b8677f-b80b-4e11-b68a-305e9dbf417f" Mar 14 09:14:25 crc kubenswrapper[4843]: I0314 09:14:25.146365 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:14:25 crc kubenswrapper[4843]: I0314 09:14:25.153045 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:14:29 crc kubenswrapper[4843]: I0314 09:14:29.781083 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 14 09:14:30 crc kubenswrapper[4843]: I0314 09:14:30.850310 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 14 09:14:32 crc kubenswrapper[4843]: I0314 09:14:32.494080 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 14 09:14:33 crc kubenswrapper[4843]: I0314 09:14:33.233612 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 14 09:14:33 crc kubenswrapper[4843]: I0314 09:14:33.412493 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 14 09:14:33 crc kubenswrapper[4843]: I0314 09:14:33.421365 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 14 09:14:33 crc kubenswrapper[4843]: I0314 09:14:33.515631 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 14 09:14:33 crc kubenswrapper[4843]: I0314 09:14:33.580209 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 14 09:14:33 crc kubenswrapper[4843]: I0314 09:14:33.723357 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 14 09:14:34 crc kubenswrapper[4843]: I0314 09:14:34.005783 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 14 09:14:34 crc kubenswrapper[4843]: I0314 09:14:34.108723 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 14 09:14:34 crc kubenswrapper[4843]: I0314 09:14:34.399100 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 14 09:14:34 crc kubenswrapper[4843]: I0314 09:14:34.405603 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 14 09:14:34 crc kubenswrapper[4843]: I0314 09:14:34.771242 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 14 09:14:34 crc kubenswrapper[4843]: I0314 09:14:34.875916 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.035158 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.041797 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.043769 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.391083 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.513224 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.532565 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.765808 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 14 09:14:35 crc kubenswrapper[4843]: I0314 09:14:35.987945 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.012794 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.167935 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.422605 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.485908 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.623534 4843 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.647341 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.744214 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.805942 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.862460 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.902576 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 14 09:14:36 crc kubenswrapper[4843]: I0314 09:14:36.931097 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.019965 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.125207 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.289085 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.314315 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.391472 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.392574 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.505671 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.665245 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.812056 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.857053 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.861206 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.883513 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.958535 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 14 09:14:37 crc kubenswrapper[4843]: I0314 09:14:37.966335 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.105522 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.226311 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.241879 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.288852 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.338428 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.360984 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.431653 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.446197 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.498724 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.655716 4843 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.686901 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.710886 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.714077 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.835383 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 14 09:14:38 crc kubenswrapper[4843]: I0314 09:14:38.895389 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.031408 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.069744 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.076023 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.110193 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.153897 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.227551 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.240398 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.337730 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.354913 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.372413 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.403112 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.444713 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.540060 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.543026 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.617838 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.621313 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.667359 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.680983 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.681082 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.790832 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 14 09:14:39 crc kubenswrapper[4843]: I0314 09:14:39.831920 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.045160 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.134152 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.248445 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.316339 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.367361 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.376645 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.395972 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.420324 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.432144 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.522681 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.646119 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.650905 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.734727 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.960220 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 14 09:14:40 crc kubenswrapper[4843]: I0314 09:14:40.961006 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.009583 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.070151 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.217790 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.224538 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.243764 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.303614 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.305927 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.382802 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.411568 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.444431 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.480935 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.499550 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.519772 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.651595 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.752051 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.833639 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.932985 4843 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 14 09:14:41 crc kubenswrapper[4843]: I0314 09:14:41.984265 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.037356 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.076872 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.223586 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.253726 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.312028 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.313465 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.324397 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.328730 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.394621 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.406164 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.458881 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.498555 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.590180 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.643901 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.697302 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.838115 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.871398 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.953132 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 14 09:14:42 crc kubenswrapper[4843]: I0314 09:14:42.960481 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.085138 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.109937 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.156553 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.168122 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.416765 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.425852 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.451723 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.453642 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.484092 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.493943 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.527836 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.531771 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.549690 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.571012 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.595976 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.603742 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.683597 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.686176 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.705567 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.707046 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.812862 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.825523 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.836382 4843 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.837082 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.889260 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.917809 4843 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.925221 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.925419 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 14 09:14:43 crc kubenswrapper[4843]: I0314 09:14:43.952836 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.004363 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.059304 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.234577 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.285849 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.290513 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.304026 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.484704 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.492168 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.512849 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.517890 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.554715 4843 ???:1] "http: TLS handshake error from 192.168.126.11:44140: no serving certificate available for the kubelet" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.590235 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.729007 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.907715 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.963538 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.976775 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 14 09:14:44 crc kubenswrapper[4843]: I0314 09:14:44.994665 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.129685 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.191221 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.198705 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.284603 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.342789 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.387733 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.387811 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.466108 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.509694 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.530902 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.604871 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.625528 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.747066 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.837899 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 14 09:14:45 crc kubenswrapper[4843]: I0314 09:14:45.934778 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.111704 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.259574 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.267725 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.268741 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.270969 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.290459 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.291411 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.377648 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.409854 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.504814 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.622054 4843 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.627457 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.627522 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.633258 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.643851 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.643801309 podStartE2EDuration="24.643801309s" podCreationTimestamp="2026-03-14 09:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:14:46.642375907 +0000 UTC m=+253.954987055" watchObservedRunningTime="2026-03-14 09:14:46.643801309 +0000 UTC m=+253.956412457" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.650982 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.713690 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.719252 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.785144 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.863544 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.928189 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.960820 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 14 09:14:46 crc kubenswrapper[4843]: I0314 09:14:46.993970 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.043330 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.048602 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.052561 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.062362 4843 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.069101 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.098843 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.315953 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.384799 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.456518 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.535225 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.596838 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.603345 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.768627 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.842976 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 14 09:14:47 crc kubenswrapper[4843]: I0314 09:14:47.969088 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.014364 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.141169 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.151909 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.296173 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.554766 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.782755 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.794701 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 14 09:14:48 crc kubenswrapper[4843]: I0314 09:14:48.910136 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.052212 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.062703 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.079742 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.197764 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.304649 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.342149 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.422024 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.675826 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 14 09:14:49 crc kubenswrapper[4843]: I0314 09:14:49.881799 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 14 09:14:50 crc kubenswrapper[4843]: I0314 09:14:50.003152 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 14 09:14:50 crc kubenswrapper[4843]: I0314 09:14:50.192983 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 14 09:14:50 crc kubenswrapper[4843]: I0314 09:14:50.193097 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 14 09:14:50 crc kubenswrapper[4843]: I0314 09:14:50.196458 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 14 09:14:50 crc kubenswrapper[4843]: I0314 09:14:50.836886 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 14 09:14:51 crc kubenswrapper[4843]: I0314 09:14:51.098813 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 14 09:14:51 crc kubenswrapper[4843]: I0314 09:14:51.154248 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 14 09:14:51 crc kubenswrapper[4843]: I0314 09:14:51.195797 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 14 09:14:51 crc kubenswrapper[4843]: I0314 09:14:51.281065 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 14 09:14:55 crc kubenswrapper[4843]: I0314 09:14:55.970082 4843 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 14 09:14:55 crc kubenswrapper[4843]: I0314 09:14:55.970870 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://84e0d02f7ef73b46bd1c83fa8e8b48f8e8fc29051d90a3b3edd8ea1d71a76130" gracePeriod=5 Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.140647 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz"] Mar 14 09:15:00 crc kubenswrapper[4843]: E0314 09:15:00.141320 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" containerName="installer" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.141343 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" containerName="installer" Mar 14 09:15:00 crc kubenswrapper[4843]: E0314 09:15:00.141371 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" containerName="oc" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.141383 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" containerName="oc" Mar 14 09:15:00 crc kubenswrapper[4843]: E0314 09:15:00.141400 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.141413 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.141611 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" containerName="oc" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.141630 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6032032-cdfd-41cd-8783-04efb653d09f" containerName="installer" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.141648 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.153360 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz"] Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.153523 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.159565 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.160000 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.268024 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5627cd7c-aab3-40a1-b844-9c4d5fba908f-config-volume\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.268199 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt8n4\" (UniqueName: \"kubernetes.io/projected/5627cd7c-aab3-40a1-b844-9c4d5fba908f-kube-api-access-pt8n4\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.268297 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5627cd7c-aab3-40a1-b844-9c4d5fba908f-secret-volume\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.370480 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5627cd7c-aab3-40a1-b844-9c4d5fba908f-config-volume\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.370613 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt8n4\" (UniqueName: \"kubernetes.io/projected/5627cd7c-aab3-40a1-b844-9c4d5fba908f-kube-api-access-pt8n4\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.370672 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5627cd7c-aab3-40a1-b844-9c4d5fba908f-secret-volume\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.373194 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5627cd7c-aab3-40a1-b844-9c4d5fba908f-config-volume\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.382100 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5627cd7c-aab3-40a1-b844-9c4d5fba908f-secret-volume\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.403921 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt8n4\" (UniqueName: \"kubernetes.io/projected/5627cd7c-aab3-40a1-b844-9c4d5fba908f-kube-api-access-pt8n4\") pod \"collect-profiles-29557995-sd7tz\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.485228 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:00 crc kubenswrapper[4843]: I0314 09:15:00.978441 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz"] Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.281208 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.282006 4843 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="84e0d02f7ef73b46bd1c83fa8e8b48f8e8fc29051d90a3b3edd8ea1d71a76130" exitCode=137 Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.284477 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" event={"ID":"5627cd7c-aab3-40a1-b844-9c4d5fba908f","Type":"ContainerStarted","Data":"957acd9fa549ac74b16cbd40d39d256c0886dddf20782ba9cb8142060f6adc9d"} Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.284577 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" event={"ID":"5627cd7c-aab3-40a1-b844-9c4d5fba908f","Type":"ContainerStarted","Data":"2b5399bdce7149489f8434bd5a8526e61e87cbcb611e8a41cc3a64ef05b7ae7c"} Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.317341 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" podStartSLOduration=1.317305264 podStartE2EDuration="1.317305264s" podCreationTimestamp="2026-03-14 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:15:01.308799468 +0000 UTC m=+268.621410596" watchObservedRunningTime="2026-03-14 09:15:01.317305264 +0000 UTC m=+268.629916432" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.567581 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.568054 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688583 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688699 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688720 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688759 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688786 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688820 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688902 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688926 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.688998 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.689161 4843 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.689185 4843 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.689196 4843 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.689207 4843 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.701353 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:15:01 crc kubenswrapper[4843]: I0314 09:15:01.791077 4843 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:02 crc kubenswrapper[4843]: I0314 09:15:02.294473 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 14 09:15:02 crc kubenswrapper[4843]: I0314 09:15:02.294618 4843 scope.go:117] "RemoveContainer" containerID="84e0d02f7ef73b46bd1c83fa8e8b48f8e8fc29051d90a3b3edd8ea1d71a76130" Mar 14 09:15:02 crc kubenswrapper[4843]: I0314 09:15:02.294733 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 14 09:15:02 crc kubenswrapper[4843]: I0314 09:15:02.297051 4843 generic.go:334] "Generic (PLEG): container finished" podID="5627cd7c-aab3-40a1-b844-9c4d5fba908f" containerID="957acd9fa549ac74b16cbd40d39d256c0886dddf20782ba9cb8142060f6adc9d" exitCode=0 Mar 14 09:15:02 crc kubenswrapper[4843]: I0314 09:15:02.297108 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" event={"ID":"5627cd7c-aab3-40a1-b844-9c4d5fba908f","Type":"ContainerDied","Data":"957acd9fa549ac74b16cbd40d39d256c0886dddf20782ba9cb8142060f6adc9d"} Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.351219 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.603517 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.720002 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5627cd7c-aab3-40a1-b844-9c4d5fba908f-config-volume\") pod \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.720090 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt8n4\" (UniqueName: \"kubernetes.io/projected/5627cd7c-aab3-40a1-b844-9c4d5fba908f-kube-api-access-pt8n4\") pod \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.720115 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5627cd7c-aab3-40a1-b844-9c4d5fba908f-secret-volume\") pod \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\" (UID: \"5627cd7c-aab3-40a1-b844-9c4d5fba908f\") " Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.721608 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5627cd7c-aab3-40a1-b844-9c4d5fba908f-config-volume" (OuterVolumeSpecName: "config-volume") pod "5627cd7c-aab3-40a1-b844-9c4d5fba908f" (UID: "5627cd7c-aab3-40a1-b844-9c4d5fba908f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.725882 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5627cd7c-aab3-40a1-b844-9c4d5fba908f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5627cd7c-aab3-40a1-b844-9c4d5fba908f" (UID: "5627cd7c-aab3-40a1-b844-9c4d5fba908f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.726506 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5627cd7c-aab3-40a1-b844-9c4d5fba908f-kube-api-access-pt8n4" (OuterVolumeSpecName: "kube-api-access-pt8n4") pod "5627cd7c-aab3-40a1-b844-9c4d5fba908f" (UID: "5627cd7c-aab3-40a1-b844-9c4d5fba908f"). InnerVolumeSpecName "kube-api-access-pt8n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.821686 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt8n4\" (UniqueName: \"kubernetes.io/projected/5627cd7c-aab3-40a1-b844-9c4d5fba908f-kube-api-access-pt8n4\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.821743 4843 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5627cd7c-aab3-40a1-b844-9c4d5fba908f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:03 crc kubenswrapper[4843]: I0314 09:15:03.821767 4843 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5627cd7c-aab3-40a1-b844-9c4d5fba908f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:04 crc kubenswrapper[4843]: I0314 09:15:04.313533 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" event={"ID":"5627cd7c-aab3-40a1-b844-9c4d5fba908f","Type":"ContainerDied","Data":"2b5399bdce7149489f8434bd5a8526e61e87cbcb611e8a41cc3a64ef05b7ae7c"} Mar 14 09:15:04 crc kubenswrapper[4843]: I0314 09:15:04.313581 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b5399bdce7149489f8434bd5a8526e61e87cbcb611e8a41cc3a64ef05b7ae7c" Mar 14 09:15:04 crc kubenswrapper[4843]: I0314 09:15:04.313611 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz" Mar 14 09:15:07 crc kubenswrapper[4843]: I0314 09:15:07.345489 4843 generic.go:334] "Generic (PLEG): container finished" podID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerID="98cb6d370b35bba7252e559ee327df9e2cac39087e0bfac591d32a12ba827dd6" exitCode=0 Mar 14 09:15:07 crc kubenswrapper[4843]: I0314 09:15:07.347442 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" event={"ID":"991014d6-3d2c-4291-91ef-db919b1c7d1c","Type":"ContainerDied","Data":"98cb6d370b35bba7252e559ee327df9e2cac39087e0bfac591d32a12ba827dd6"} Mar 14 09:15:07 crc kubenswrapper[4843]: I0314 09:15:07.349687 4843 scope.go:117] "RemoveContainer" containerID="98cb6d370b35bba7252e559ee327df9e2cac39087e0bfac591d32a12ba827dd6" Mar 14 09:15:08 crc kubenswrapper[4843]: I0314 09:15:08.356944 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" event={"ID":"991014d6-3d2c-4291-91ef-db919b1c7d1c","Type":"ContainerStarted","Data":"516e46ec7d51734a38b7cb27a7aada007663c7b7c8e15496b98ef5ac5de39984"} Mar 14 09:15:08 crc kubenswrapper[4843]: I0314 09:15:08.357918 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:15:08 crc kubenswrapper[4843]: I0314 09:15:08.360491 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:15:09 crc kubenswrapper[4843]: I0314 09:15:09.610354 4843 ???:1] "http: TLS handshake error from 192.168.126.11:55192: no serving certificate available for the kubelet" Mar 14 09:15:15 crc kubenswrapper[4843]: I0314 09:15:15.387729 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:15:15 crc kubenswrapper[4843]: I0314 09:15:15.388498 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:15:15 crc kubenswrapper[4843]: I0314 09:15:15.388561 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:15:15 crc kubenswrapper[4843]: I0314 09:15:15.389338 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1eec10ad2604b241d08a4efa3556d36326f8f05b555f3f2dc5588bc1ea55449f"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:15:15 crc kubenswrapper[4843]: I0314 09:15:15.389456 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://1eec10ad2604b241d08a4efa3556d36326f8f05b555f3f2dc5588bc1ea55449f" gracePeriod=600 Mar 14 09:15:16 crc kubenswrapper[4843]: I0314 09:15:16.405867 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="1eec10ad2604b241d08a4efa3556d36326f8f05b555f3f2dc5588bc1ea55449f" exitCode=0 Mar 14 09:15:16 crc kubenswrapper[4843]: I0314 09:15:16.405980 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"1eec10ad2604b241d08a4efa3556d36326f8f05b555f3f2dc5588bc1ea55449f"} Mar 14 09:15:16 crc kubenswrapper[4843]: I0314 09:15:16.406265 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"bb6c919a4c034719ae1f3b02fb3d4bb9d30d21fd25a1a82acf24c004a65a3907"} Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.798332 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mss9d"] Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.799166 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mss9d" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="registry-server" containerID="cri-o://1d2523b4024629aa058e87c4903c05bb59ed654faa17d84400b74924e537fb45" gracePeriod=30 Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.811412 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kzj4r"] Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.811936 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kzj4r" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="registry-server" containerID="cri-o://2bf4eb7984dd563980a00e89f566f8f2e609c773ddfe82f9489a601032fa90e9" gracePeriod=30 Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.819633 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xtq2f"] Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.819866 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" containerID="cri-o://516e46ec7d51734a38b7cb27a7aada007663c7b7c8e15496b98ef5ac5de39984" gracePeriod=30 Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.840072 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmpq6"] Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.840380 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mmpq6" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="registry-server" containerID="cri-o://a32ae428a788540ef28f5251d0de6014324adbc9ee5b3deb8cf76358dfda7ab6" gracePeriod=30 Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.846638 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jfzjg"] Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.846885 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jfzjg" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="registry-server" containerID="cri-o://25bf24975d63a934c63c123325ef1ca1d58ee2581e50f26a6ffce3fecf7b9cb3" gracePeriod=30 Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.865852 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-565tx"] Mar 14 09:15:53 crc kubenswrapper[4843]: E0314 09:15:53.866124 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5627cd7c-aab3-40a1-b844-9c4d5fba908f" containerName="collect-profiles" Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.866139 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5627cd7c-aab3-40a1-b844-9c4d5fba908f" containerName="collect-profiles" Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.866242 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5627cd7c-aab3-40a1-b844-9c4d5fba908f" containerName="collect-profiles" Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.866695 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.882103 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-565tx"] Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.997985 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3bd512f7-d36b-43b0-94af-cee8982d7630-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.998046 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbxt9\" (UniqueName: \"kubernetes.io/projected/3bd512f7-d36b-43b0-94af-cee8982d7630-kube-api-access-bbxt9\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:53 crc kubenswrapper[4843]: I0314 09:15:53.998258 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3bd512f7-d36b-43b0-94af-cee8982d7630-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.099714 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3bd512f7-d36b-43b0-94af-cee8982d7630-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.099783 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbxt9\" (UniqueName: \"kubernetes.io/projected/3bd512f7-d36b-43b0-94af-cee8982d7630-kube-api-access-bbxt9\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.099853 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3bd512f7-d36b-43b0-94af-cee8982d7630-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.101117 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3bd512f7-d36b-43b0-94af-cee8982d7630-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.120867 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbxt9\" (UniqueName: \"kubernetes.io/projected/3bd512f7-d36b-43b0-94af-cee8982d7630-kube-api-access-bbxt9\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.121432 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3bd512f7-d36b-43b0-94af-cee8982d7630-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-565tx\" (UID: \"3bd512f7-d36b-43b0-94af-cee8982d7630\") " pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.207940 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.623732 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-565tx"] Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.646000 4843 generic.go:334] "Generic (PLEG): container finished" podID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerID="2bf4eb7984dd563980a00e89f566f8f2e609c773ddfe82f9489a601032fa90e9" exitCode=0 Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.646071 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzj4r" event={"ID":"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f","Type":"ContainerDied","Data":"2bf4eb7984dd563980a00e89f566f8f2e609c773ddfe82f9489a601032fa90e9"} Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.651709 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerID="1d2523b4024629aa058e87c4903c05bb59ed654faa17d84400b74924e537fb45" exitCode=0 Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.651800 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mss9d" event={"ID":"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb","Type":"ContainerDied","Data":"1d2523b4024629aa058e87c4903c05bb59ed654faa17d84400b74924e537fb45"} Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.653587 4843 generic.go:334] "Generic (PLEG): container finished" podID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerID="516e46ec7d51734a38b7cb27a7aada007663c7b7c8e15496b98ef5ac5de39984" exitCode=0 Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.653629 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" event={"ID":"991014d6-3d2c-4291-91ef-db919b1c7d1c","Type":"ContainerDied","Data":"516e46ec7d51734a38b7cb27a7aada007663c7b7c8e15496b98ef5ac5de39984"} Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.653668 4843 scope.go:117] "RemoveContainer" containerID="98cb6d370b35bba7252e559ee327df9e2cac39087e0bfac591d32a12ba827dd6" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.701761 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.813559 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g98sp\" (UniqueName: \"kubernetes.io/projected/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-kube-api-access-g98sp\") pod \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.813715 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-catalog-content\") pod \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.813736 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-utilities\") pod \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\" (UID: \"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb\") " Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.822878 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-utilities" (OuterVolumeSpecName: "utilities") pod "0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" (UID: "0c1ae1f4-158a-4e77-ae1f-1f5db661bafb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.824233 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-kube-api-access-g98sp" (OuterVolumeSpecName: "kube-api-access-g98sp") pod "0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" (UID: "0c1ae1f4-158a-4e77-ae1f-1f5db661bafb"). InnerVolumeSpecName "kube-api-access-g98sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.841050 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.875377 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" (UID: "0c1ae1f4-158a-4e77-ae1f-1f5db661bafb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.915416 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.915447 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:54 crc kubenswrapper[4843]: I0314 09:15:54.915459 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g98sp\" (UniqueName: \"kubernetes.io/projected/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb-kube-api-access-g98sp\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.016068 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-catalog-content\") pod \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.016188 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhsjs\" (UniqueName: \"kubernetes.io/projected/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-kube-api-access-xhsjs\") pod \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.016214 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-utilities\") pod \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\" (UID: \"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f\") " Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.017057 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-utilities" (OuterVolumeSpecName: "utilities") pod "076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" (UID: "076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.019760 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-kube-api-access-xhsjs" (OuterVolumeSpecName: "kube-api-access-xhsjs") pod "076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" (UID: "076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f"). InnerVolumeSpecName "kube-api-access-xhsjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.069564 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" (UID: "076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.117803 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhsjs\" (UniqueName: \"kubernetes.io/projected/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-kube-api-access-xhsjs\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.117831 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.117840 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.681911 4843 generic.go:334] "Generic (PLEG): container finished" podID="8220c353-b74c-4703-ac79-af96379e89fe" containerID="a32ae428a788540ef28f5251d0de6014324adbc9ee5b3deb8cf76358dfda7ab6" exitCode=0 Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.682021 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmpq6" event={"ID":"8220c353-b74c-4703-ac79-af96379e89fe","Type":"ContainerDied","Data":"a32ae428a788540ef28f5251d0de6014324adbc9ee5b3deb8cf76358dfda7ab6"} Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.687815 4843 generic.go:334] "Generic (PLEG): container finished" podID="3568e895-1a84-487a-9042-376d2bb9e28f" containerID="25bf24975d63a934c63c123325ef1ca1d58ee2581e50f26a6ffce3fecf7b9cb3" exitCode=0 Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.687884 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerDied","Data":"25bf24975d63a934c63c123325ef1ca1d58ee2581e50f26a6ffce3fecf7b9cb3"} Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.688920 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" event={"ID":"3bd512f7-d36b-43b0-94af-cee8982d7630","Type":"ContainerStarted","Data":"73397649f9c1cb5a5c8e9b32f8f4ca2bad9129cde43aee24121eb2649154c825"} Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.691196 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzj4r" event={"ID":"076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f","Type":"ContainerDied","Data":"72e753957c21bdc806b07b295fe82dc111a25d82d3e3235682af56fcd1d612dd"} Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.691242 4843 scope.go:117] "RemoveContainer" containerID="2bf4eb7984dd563980a00e89f566f8f2e609c773ddfe82f9489a601032fa90e9" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.691386 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzj4r" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.697615 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mss9d" event={"ID":"0c1ae1f4-158a-4e77-ae1f-1f5db661bafb","Type":"ContainerDied","Data":"7ef4eb91812527ad7654bccd71811449a0a07f27f2faf65e8a1f28b792f0b475"} Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.697713 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mss9d" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.713042 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kzj4r"] Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.721499 4843 scope.go:117] "RemoveContainer" containerID="92370561e6ec03fda689fe0ed7da1c04a650f66083a411f53f9c4b074c614602" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.725501 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kzj4r"] Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.734390 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mss9d"] Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.739639 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mss9d"] Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.757474 4843 scope.go:117] "RemoveContainer" containerID="2a209684e8e5c6b45db759e4ffa84a740aabdbcf8d9effb6c818e1694d55efc0" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.780666 4843 scope.go:117] "RemoveContainer" containerID="1d2523b4024629aa058e87c4903c05bb59ed654faa17d84400b74924e537fb45" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.800575 4843 scope.go:117] "RemoveContainer" containerID="b2b7a85bc0131426f55ffe652dd79762a2397f02cb671b3cecd668d347487dbe" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.821046 4843 scope.go:117] "RemoveContainer" containerID="bc486ebf3271c5a67c24e45698b89cc184cfd9d386e70c7553e094393635cfeb" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.836454 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.926991 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-catalog-content\") pod \"3568e895-1a84-487a-9042-376d2bb9e28f\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.927107 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5z9q\" (UniqueName: \"kubernetes.io/projected/3568e895-1a84-487a-9042-376d2bb9e28f-kube-api-access-t5z9q\") pod \"3568e895-1a84-487a-9042-376d2bb9e28f\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.927157 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-utilities\") pod \"3568e895-1a84-487a-9042-376d2bb9e28f\" (UID: \"3568e895-1a84-487a-9042-376d2bb9e28f\") " Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.928095 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-utilities" (OuterVolumeSpecName: "utilities") pod "3568e895-1a84-487a-9042-376d2bb9e28f" (UID: "3568e895-1a84-487a-9042-376d2bb9e28f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:55 crc kubenswrapper[4843]: I0314 09:15:55.935867 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3568e895-1a84-487a-9042-376d2bb9e28f-kube-api-access-t5z9q" (OuterVolumeSpecName: "kube-api-access-t5z9q") pod "3568e895-1a84-487a-9042-376d2bb9e28f" (UID: "3568e895-1a84-487a-9042-376d2bb9e28f"). InnerVolumeSpecName "kube-api-access-t5z9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.029583 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5z9q\" (UniqueName: \"kubernetes.io/projected/3568e895-1a84-487a-9042-376d2bb9e28f-kube-api-access-t5z9q\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.029629 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.058153 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3568e895-1a84-487a-9042-376d2bb9e28f" (UID: "3568e895-1a84-487a-9042-376d2bb9e28f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.130381 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3568e895-1a84-487a-9042-376d2bb9e28f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.253817 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.277478 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.332890 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pnk2\" (UniqueName: \"kubernetes.io/projected/991014d6-3d2c-4291-91ef-db919b1c7d1c-kube-api-access-9pnk2\") pod \"991014d6-3d2c-4291-91ef-db919b1c7d1c\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.332949 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-operator-metrics\") pod \"991014d6-3d2c-4291-91ef-db919b1c7d1c\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.333022 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-trusted-ca\") pod \"991014d6-3d2c-4291-91ef-db919b1c7d1c\" (UID: \"991014d6-3d2c-4291-91ef-db919b1c7d1c\") " Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.334204 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "991014d6-3d2c-4291-91ef-db919b1c7d1c" (UID: "991014d6-3d2c-4291-91ef-db919b1c7d1c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.341921 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/991014d6-3d2c-4291-91ef-db919b1c7d1c-kube-api-access-9pnk2" (OuterVolumeSpecName: "kube-api-access-9pnk2") pod "991014d6-3d2c-4291-91ef-db919b1c7d1c" (UID: "991014d6-3d2c-4291-91ef-db919b1c7d1c"). InnerVolumeSpecName "kube-api-access-9pnk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.342117 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "991014d6-3d2c-4291-91ef-db919b1c7d1c" (UID: "991014d6-3d2c-4291-91ef-db919b1c7d1c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.434648 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-utilities\") pod \"8220c353-b74c-4703-ac79-af96379e89fe\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.434803 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmtt4\" (UniqueName: \"kubernetes.io/projected/8220c353-b74c-4703-ac79-af96379e89fe-kube-api-access-gmtt4\") pod \"8220c353-b74c-4703-ac79-af96379e89fe\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.434913 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-catalog-content\") pod \"8220c353-b74c-4703-ac79-af96379e89fe\" (UID: \"8220c353-b74c-4703-ac79-af96379e89fe\") " Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.435907 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-utilities" (OuterVolumeSpecName: "utilities") pod "8220c353-b74c-4703-ac79-af96379e89fe" (UID: "8220c353-b74c-4703-ac79-af96379e89fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.436493 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pnk2\" (UniqueName: \"kubernetes.io/projected/991014d6-3d2c-4291-91ef-db919b1c7d1c-kube-api-access-9pnk2\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.436521 4843 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.436535 4843 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/991014d6-3d2c-4291-91ef-db919b1c7d1c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.439997 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8220c353-b74c-4703-ac79-af96379e89fe-kube-api-access-gmtt4" (OuterVolumeSpecName: "kube-api-access-gmtt4") pod "8220c353-b74c-4703-ac79-af96379e89fe" (UID: "8220c353-b74c-4703-ac79-af96379e89fe"). InnerVolumeSpecName "kube-api-access-gmtt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.478103 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8220c353-b74c-4703-ac79-af96379e89fe" (UID: "8220c353-b74c-4703-ac79-af96379e89fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.537899 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.537942 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmtt4\" (UniqueName: \"kubernetes.io/projected/8220c353-b74c-4703-ac79-af96379e89fe-kube-api-access-gmtt4\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.537956 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8220c353-b74c-4703-ac79-af96379e89fe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.705546 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" event={"ID":"991014d6-3d2c-4291-91ef-db919b1c7d1c","Type":"ContainerDied","Data":"dc871c860e5aa687a79991225f75d3a6ab292336b24b55db833f6b9f7f9e3031"} Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.705558 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xtq2f" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.705606 4843 scope.go:117] "RemoveContainer" containerID="516e46ec7d51734a38b7cb27a7aada007663c7b7c8e15496b98ef5ac5de39984" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.709989 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfzjg" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.709991 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfzjg" event={"ID":"3568e895-1a84-487a-9042-376d2bb9e28f","Type":"ContainerDied","Data":"58ee98669e99b6063d4a255923e898688b4bb00a48cbdd166a48c5976a1ea9f0"} Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.712085 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" event={"ID":"3bd512f7-d36b-43b0-94af-cee8982d7630","Type":"ContainerStarted","Data":"869d2a935ac15be074e35c2351a63dea8c3f1b80a1f25d64573f32e9cac67e19"} Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.712315 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.718120 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.721722 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mmpq6" event={"ID":"8220c353-b74c-4703-ac79-af96379e89fe","Type":"ContainerDied","Data":"f8b73f6dcdbb225316a081ed59715469f3075e11b49b7eb310bc589582ca274c"} Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.721812 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mmpq6" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.736478 4843 scope.go:117] "RemoveContainer" containerID="25bf24975d63a934c63c123325ef1ca1d58ee2581e50f26a6ffce3fecf7b9cb3" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.741554 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-565tx" podStartSLOduration=3.7415341079999997 podStartE2EDuration="3.741534108s" podCreationTimestamp="2026-03-14 09:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:15:56.736528401 +0000 UTC m=+324.049139529" watchObservedRunningTime="2026-03-14 09:15:56.741534108 +0000 UTC m=+324.054145236" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.758034 4843 scope.go:117] "RemoveContainer" containerID="57cf099bf0adae4642557f0162a505ab0a88e5ce692b512a1a0fa6d5189eb8e8" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.758890 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jfzjg"] Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.764124 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jfzjg"] Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.779410 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xtq2f"] Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.782826 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xtq2f"] Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.785598 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmpq6"] Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.788101 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mmpq6"] Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.794219 4843 scope.go:117] "RemoveContainer" containerID="ca20868760ef13e0cc0ebee4f5469dd4fa4773c40ffc5c993defc2b36e97e34d" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.847759 4843 scope.go:117] "RemoveContainer" containerID="a32ae428a788540ef28f5251d0de6014324adbc9ee5b3deb8cf76358dfda7ab6" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.863672 4843 scope.go:117] "RemoveContainer" containerID="c407df7c6a78977956bb2212f97a460d5af3d7abb779e5d102cad33b1272e18f" Mar 14 09:15:56 crc kubenswrapper[4843]: I0314 09:15:56.882025 4843 scope.go:117] "RemoveContainer" containerID="035fd8923dba17fd8d08379cd43918a0f3ac4aa51fc14f6e35af74b8e9d9978a" Mar 14 09:15:57 crc kubenswrapper[4843]: I0314 09:15:57.345664 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" path="/var/lib/kubelet/pods/076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f/volumes" Mar 14 09:15:57 crc kubenswrapper[4843]: I0314 09:15:57.346797 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" path="/var/lib/kubelet/pods/0c1ae1f4-158a-4e77-ae1f-1f5db661bafb/volumes" Mar 14 09:15:57 crc kubenswrapper[4843]: I0314 09:15:57.347384 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" path="/var/lib/kubelet/pods/3568e895-1a84-487a-9042-376d2bb9e28f/volumes" Mar 14 09:15:57 crc kubenswrapper[4843]: I0314 09:15:57.348410 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8220c353-b74c-4703-ac79-af96379e89fe" path="/var/lib/kubelet/pods/8220c353-b74c-4703-ac79-af96379e89fe/volumes" Mar 14 09:15:57 crc kubenswrapper[4843]: I0314 09:15:57.348986 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" path="/var/lib/kubelet/pods/991014d6-3d2c-4291-91ef-db919b1c7d1c/volumes" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.021992 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tt7cz"] Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.022786 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.022876 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.022933 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.022994 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023053 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023104 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023156 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023213 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023264 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023345 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023402 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023453 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023502 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023552 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023606 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023660 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023712 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023761 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023812 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023868 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.023924 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.023978 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.024032 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024087 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.024148 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024203 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="extract-utilities" Mar 14 09:15:58 crc kubenswrapper[4843]: E0314 09:15:58.024254 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024333 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="extract-content" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024479 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3568e895-1a84-487a-9042-376d2bb9e28f" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024557 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024627 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="991014d6-3d2c-4291-91ef-db919b1c7d1c" containerName="marketplace-operator" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024681 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1ae1f4-158a-4e77-ae1f-1f5db661bafb" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024737 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8220c353-b74c-4703-ac79-af96379e89fe" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.024796 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="076caeb9-d7f0-4ff9-90dd-ae912b7f4a9f" containerName="registry-server" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.025874 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.028945 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.034239 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt7cz"] Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.158539 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4j4l\" (UniqueName: \"kubernetes.io/projected/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-kube-api-access-v4j4l\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.158600 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-catalog-content\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.158629 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-utilities\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.199035 4843 csr.go:261] certificate signing request csr-sk64q is approved, waiting to be issued Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.220041 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-llm5h"] Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.221012 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.225220 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.225544 4843 csr.go:257] certificate signing request csr-sk64q is issued Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.259671 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-utilities\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.259792 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4j4l\" (UniqueName: \"kubernetes.io/projected/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-kube-api-access-v4j4l\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.259837 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-catalog-content\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.260172 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-utilities\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.260190 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-catalog-content\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.266523 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llm5h"] Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.287074 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4j4l\" (UniqueName: \"kubernetes.io/projected/38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6-kube-api-access-v4j4l\") pod \"certified-operators-tt7cz\" (UID: \"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6\") " pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.351837 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.361184 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/353b52a0-aebf-426a-9006-74c1d24e8c48-utilities\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.361334 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgqnb\" (UniqueName: \"kubernetes.io/projected/353b52a0-aebf-426a-9006-74c1d24e8c48-kube-api-access-vgqnb\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.361430 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/353b52a0-aebf-426a-9006-74c1d24e8c48-catalog-content\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.464173 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/353b52a0-aebf-426a-9006-74c1d24e8c48-utilities\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.464537 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgqnb\" (UniqueName: \"kubernetes.io/projected/353b52a0-aebf-426a-9006-74c1d24e8c48-kube-api-access-vgqnb\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.464580 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/353b52a0-aebf-426a-9006-74c1d24e8c48-catalog-content\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.464691 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/353b52a0-aebf-426a-9006-74c1d24e8c48-utilities\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.464958 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/353b52a0-aebf-426a-9006-74c1d24e8c48-catalog-content\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.486533 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgqnb\" (UniqueName: \"kubernetes.io/projected/353b52a0-aebf-426a-9006-74c1d24e8c48-kube-api-access-vgqnb\") pod \"community-operators-llm5h\" (UID: \"353b52a0-aebf-426a-9006-74c1d24e8c48\") " pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.551035 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.748975 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt7cz"] Mar 14 09:15:58 crc kubenswrapper[4843]: W0314 09:15:58.753729 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38ba05ff_31cc_4fa9_a6fc_e374e8c6ede6.slice/crio-4399f132033e0b4af1f2b5b50cf41346f6eb450cb64434c3d79e28dc40e59c41 WatchSource:0}: Error finding container 4399f132033e0b4af1f2b5b50cf41346f6eb450cb64434c3d79e28dc40e59c41: Status 404 returned error can't find the container with id 4399f132033e0b4af1f2b5b50cf41346f6eb450cb64434c3d79e28dc40e59c41 Mar 14 09:15:58 crc kubenswrapper[4843]: I0314 09:15:58.954454 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-llm5h"] Mar 14 09:15:58 crc kubenswrapper[4843]: W0314 09:15:58.965781 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod353b52a0_aebf_426a_9006_74c1d24e8c48.slice/crio-d7b4dfd93fc4c9d549099d0d412ec60ee6c99907f179781d67ab38a9aa8729bc WatchSource:0}: Error finding container d7b4dfd93fc4c9d549099d0d412ec60ee6c99907f179781d67ab38a9aa8729bc: Status 404 returned error can't find the container with id d7b4dfd93fc4c9d549099d0d412ec60ee6c99907f179781d67ab38a9aa8729bc Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.227824 4843 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-18 14:51:49.485273801 +0000 UTC Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.227880 4843 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5981h35m50.257398709s for next certificate rotation Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.748098 4843 generic.go:334] "Generic (PLEG): container finished" podID="38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6" containerID="61b0dbe738bf491e755b7c0df08e358c05c7109b2e04ab0068923c534f80ad7f" exitCode=0 Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.748149 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt7cz" event={"ID":"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6","Type":"ContainerDied","Data":"61b0dbe738bf491e755b7c0df08e358c05c7109b2e04ab0068923c534f80ad7f"} Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.748456 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt7cz" event={"ID":"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6","Type":"ContainerStarted","Data":"4399f132033e0b4af1f2b5b50cf41346f6eb450cb64434c3d79e28dc40e59c41"} Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.749815 4843 generic.go:334] "Generic (PLEG): container finished" podID="353b52a0-aebf-426a-9006-74c1d24e8c48" containerID="d8b614f8992f07e47ef5c1397e0f5e71b40d1fcc15c6c1e7a5d119f23de33509" exitCode=0 Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.749859 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llm5h" event={"ID":"353b52a0-aebf-426a-9006-74c1d24e8c48","Type":"ContainerDied","Data":"d8b614f8992f07e47ef5c1397e0f5e71b40d1fcc15c6c1e7a5d119f23de33509"} Mar 14 09:15:59 crc kubenswrapper[4843]: I0314 09:15:59.749891 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llm5h" event={"ID":"353b52a0-aebf-426a-9006-74c1d24e8c48","Type":"ContainerStarted","Data":"d7b4dfd93fc4c9d549099d0d412ec60ee6c99907f179781d67ab38a9aa8729bc"} Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.143683 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557996-tplp2"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.144389 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.146212 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.147112 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.147433 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.148593 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557996-tplp2"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.228912 4843 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-27 22:10:42.296609784 +0000 UTC Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.228965 4843 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6924h54m42.067648137s for next certificate rotation Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.296171 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9d7w\" (UniqueName: \"kubernetes.io/projected/1108730f-bc5a-472b-98d5-087d6e04ad2a-kube-api-access-m9d7w\") pod \"auto-csr-approver-29557996-tplp2\" (UID: \"1108730f-bc5a-472b-98d5-087d6e04ad2a\") " pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.397938 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9d7w\" (UniqueName: \"kubernetes.io/projected/1108730f-bc5a-472b-98d5-087d6e04ad2a-kube-api-access-m9d7w\") pod \"auto-csr-approver-29557996-tplp2\" (UID: \"1108730f-bc5a-472b-98d5-087d6e04ad2a\") " pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.423948 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9d7w\" (UniqueName: \"kubernetes.io/projected/1108730f-bc5a-472b-98d5-087d6e04ad2a-kube-api-access-m9d7w\") pod \"auto-csr-approver-29557996-tplp2\" (UID: \"1108730f-bc5a-472b-98d5-087d6e04ad2a\") " pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.424018 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tfbnl"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.427767 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.430824 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.441196 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfbnl"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.491831 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.601458 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/166fa22e-f961-4d02-bd2a-a6e215204f4b-utilities\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.602027 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lbx7\" (UniqueName: \"kubernetes.io/projected/166fa22e-f961-4d02-bd2a-a6e215204f4b-kube-api-access-6lbx7\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.602078 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/166fa22e-f961-4d02-bd2a-a6e215204f4b-catalog-content\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.626425 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vx4kc"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.627379 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.630984 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.664798 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vx4kc"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.704774 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lbx7\" (UniqueName: \"kubernetes.io/projected/166fa22e-f961-4d02-bd2a-a6e215204f4b-kube-api-access-6lbx7\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.704839 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/166fa22e-f961-4d02-bd2a-a6e215204f4b-catalog-content\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.704912 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/166fa22e-f961-4d02-bd2a-a6e215204f4b-utilities\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.706002 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/166fa22e-f961-4d02-bd2a-a6e215204f4b-utilities\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.706434 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/166fa22e-f961-4d02-bd2a-a6e215204f4b-catalog-content\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.722699 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lbx7\" (UniqueName: \"kubernetes.io/projected/166fa22e-f961-4d02-bd2a-a6e215204f4b-kube-api-access-6lbx7\") pod \"redhat-marketplace-tfbnl\" (UID: \"166fa22e-f961-4d02-bd2a-a6e215204f4b\") " pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.777093 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.806240 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pvqc\" (UniqueName: \"kubernetes.io/projected/f9c9423f-f5aa-4a79-8084-f775e9a8af97-kube-api-access-2pvqc\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.806322 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9c9423f-f5aa-4a79-8084-f775e9a8af97-utilities\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.806386 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9c9423f-f5aa-4a79-8084-f775e9a8af97-catalog-content\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.908186 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pvqc\" (UniqueName: \"kubernetes.io/projected/f9c9423f-f5aa-4a79-8084-f775e9a8af97-kube-api-access-2pvqc\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.908300 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9c9423f-f5aa-4a79-8084-f775e9a8af97-utilities\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.909264 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9c9423f-f5aa-4a79-8084-f775e9a8af97-catalog-content\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.909723 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9c9423f-f5aa-4a79-8084-f775e9a8af97-utilities\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.909994 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9c9423f-f5aa-4a79-8084-f775e9a8af97-catalog-content\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.931440 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557996-tplp2"] Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.945655 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pvqc\" (UniqueName: \"kubernetes.io/projected/f9c9423f-f5aa-4a79-8084-f775e9a8af97-kube-api-access-2pvqc\") pod \"redhat-operators-vx4kc\" (UID: \"f9c9423f-f5aa-4a79-8084-f775e9a8af97\") " pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:00 crc kubenswrapper[4843]: I0314 09:16:00.967958 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:01 crc kubenswrapper[4843]: I0314 09:16:01.761071 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557996-tplp2" event={"ID":"1108730f-bc5a-472b-98d5-087d6e04ad2a","Type":"ContainerStarted","Data":"442ec4ea34ee7dee690258878539a76f0e2731d06b304f11d5d339225959202c"} Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.051807 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8qk6m"] Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.052927 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.079102 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8qk6m"] Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242465 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242532 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-bound-sa-token\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242563 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242599 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-trusted-ca\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242680 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-registry-certificates\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242705 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t8bj\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-kube-api-access-5t8bj\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242764 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-registry-tls\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.242787 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.275500 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343418 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-registry-tls\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343466 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343509 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-bound-sa-token\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343532 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343570 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-trusted-ca\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343594 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-registry-certificates\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.343615 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t8bj\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-kube-api-access-5t8bj\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.344438 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.346147 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-trusted-ca\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.346318 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-registry-certificates\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.350459 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.350636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-registry-tls\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.361761 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t8bj\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-kube-api-access-5t8bj\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.370506 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aef4eb4a-915e-4768-ae8b-8c166b4fa8f2-bound-sa-token\") pod \"image-registry-66df7c8f76-8qk6m\" (UID: \"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2\") " pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:10 crc kubenswrapper[4843]: I0314 09:16:10.372092 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:11 crc kubenswrapper[4843]: I0314 09:16:11.900621 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfbnl"] Mar 14 09:16:11 crc kubenswrapper[4843]: W0314 09:16:11.912691 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod166fa22e_f961_4d02_bd2a_a6e215204f4b.slice/crio-d3e69919c8b98ac168a978c68b642144772f0e182cc1fc4248cfd51076cf7496 WatchSource:0}: Error finding container d3e69919c8b98ac168a978c68b642144772f0e182cc1fc4248cfd51076cf7496: Status 404 returned error can't find the container with id d3e69919c8b98ac168a978c68b642144772f0e182cc1fc4248cfd51076cf7496 Mar 14 09:16:11 crc kubenswrapper[4843]: I0314 09:16:11.953325 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vx4kc"] Mar 14 09:16:11 crc kubenswrapper[4843]: W0314 09:16:11.965883 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9c9423f_f5aa_4a79_8084_f775e9a8af97.slice/crio-952aff4b3e977afdc1da6d7b00df9997b392c949267cd285ae827f6b45e6419a WatchSource:0}: Error finding container 952aff4b3e977afdc1da6d7b00df9997b392c949267cd285ae827f6b45e6419a: Status 404 returned error can't find the container with id 952aff4b3e977afdc1da6d7b00df9997b392c949267cd285ae827f6b45e6419a Mar 14 09:16:11 crc kubenswrapper[4843]: I0314 09:16:11.989197 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8qk6m"] Mar 14 09:16:11 crc kubenswrapper[4843]: W0314 09:16:11.997119 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaef4eb4a_915e_4768_ae8b_8c166b4fa8f2.slice/crio-7a11cb3e542436c6e3e1e70ee7dc28eabf99182468e045e7474d261358efc128 WatchSource:0}: Error finding container 7a11cb3e542436c6e3e1e70ee7dc28eabf99182468e045e7474d261358efc128: Status 404 returned error can't find the container with id 7a11cb3e542436c6e3e1e70ee7dc28eabf99182468e045e7474d261358efc128 Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.826345 4843 generic.go:334] "Generic (PLEG): container finished" podID="38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6" containerID="e1418f3473beb9688161a6323fe0ad7bfc6b3ddcba1901cd12bfd2812de71cf2" exitCode=0 Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.826433 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt7cz" event={"ID":"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6","Type":"ContainerDied","Data":"e1418f3473beb9688161a6323fe0ad7bfc6b3ddcba1901cd12bfd2812de71cf2"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.828812 4843 generic.go:334] "Generic (PLEG): container finished" podID="1108730f-bc5a-472b-98d5-087d6e04ad2a" containerID="641bd9b39d70fdb98f7edbd95c7636b645a2fa7fdb0224d8776d9c156f24f6f8" exitCode=0 Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.828900 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557996-tplp2" event={"ID":"1108730f-bc5a-472b-98d5-087d6e04ad2a","Type":"ContainerDied","Data":"641bd9b39d70fdb98f7edbd95c7636b645a2fa7fdb0224d8776d9c156f24f6f8"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.830392 4843 generic.go:334] "Generic (PLEG): container finished" podID="166fa22e-f961-4d02-bd2a-a6e215204f4b" containerID="09116bcceff5cf03f149cf5f20559dbe9c4beb872f150f001da05a80a6fc9069" exitCode=0 Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.830464 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfbnl" event={"ID":"166fa22e-f961-4d02-bd2a-a6e215204f4b","Type":"ContainerDied","Data":"09116bcceff5cf03f149cf5f20559dbe9c4beb872f150f001da05a80a6fc9069"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.830509 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfbnl" event={"ID":"166fa22e-f961-4d02-bd2a-a6e215204f4b","Type":"ContainerStarted","Data":"d3e69919c8b98ac168a978c68b642144772f0e182cc1fc4248cfd51076cf7496"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.832873 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" event={"ID":"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2","Type":"ContainerStarted","Data":"689966b548b88e6e3c784b37a1d7ccf9285d5b3a297480672c36b6462a19b198"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.832908 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" event={"ID":"aef4eb4a-915e-4768-ae8b-8c166b4fa8f2","Type":"ContainerStarted","Data":"7a11cb3e542436c6e3e1e70ee7dc28eabf99182468e045e7474d261358efc128"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.833114 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.836457 4843 generic.go:334] "Generic (PLEG): container finished" podID="353b52a0-aebf-426a-9006-74c1d24e8c48" containerID="1eec3b00bf534cb782721c2a4c534f6483e0180df192f7a8617fec4e5a84fee6" exitCode=0 Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.836523 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llm5h" event={"ID":"353b52a0-aebf-426a-9006-74c1d24e8c48","Type":"ContainerDied","Data":"1eec3b00bf534cb782721c2a4c534f6483e0180df192f7a8617fec4e5a84fee6"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.840655 4843 generic.go:334] "Generic (PLEG): container finished" podID="f9c9423f-f5aa-4a79-8084-f775e9a8af97" containerID="42e6a88635f6bf3485d00977a2ca59c2c00be089287e6b2f405d715609199f7d" exitCode=0 Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.840706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx4kc" event={"ID":"f9c9423f-f5aa-4a79-8084-f775e9a8af97","Type":"ContainerDied","Data":"42e6a88635f6bf3485d00977a2ca59c2c00be089287e6b2f405d715609199f7d"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.840744 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx4kc" event={"ID":"f9c9423f-f5aa-4a79-8084-f775e9a8af97","Type":"ContainerStarted","Data":"952aff4b3e977afdc1da6d7b00df9997b392c949267cd285ae827f6b45e6419a"} Mar 14 09:16:12 crc kubenswrapper[4843]: I0314 09:16:12.895900 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" podStartSLOduration=2.895881494 podStartE2EDuration="2.895881494s" podCreationTimestamp="2026-03-14 09:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:16:12.891466947 +0000 UTC m=+340.204078075" watchObservedRunningTime="2026-03-14 09:16:12.895881494 +0000 UTC m=+340.208492622" Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.848485 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt7cz" event={"ID":"38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6","Type":"ContainerStarted","Data":"ba84f0cd71ecb2fe6bd66769dd75cc3ed37b33f685f3177fe99e4e9e343e986d"} Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.850539 4843 generic.go:334] "Generic (PLEG): container finished" podID="166fa22e-f961-4d02-bd2a-a6e215204f4b" containerID="c917e11cddbd2740f05e581cd08e7cfdbc3051b164eb9f30029d220e33ca14a8" exitCode=0 Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.850632 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfbnl" event={"ID":"166fa22e-f961-4d02-bd2a-a6e215204f4b","Type":"ContainerDied","Data":"c917e11cddbd2740f05e581cd08e7cfdbc3051b164eb9f30029d220e33ca14a8"} Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.853409 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-llm5h" event={"ID":"353b52a0-aebf-426a-9006-74c1d24e8c48","Type":"ContainerStarted","Data":"589f4cd9533354a26ed29f49184beead54eaa32c2d9f43ca66a747579eca1951"} Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.857008 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx4kc" event={"ID":"f9c9423f-f5aa-4a79-8084-f775e9a8af97","Type":"ContainerStarted","Data":"06736abea2e812a113a76e07c6a59d639353333628be3fe9f0ea4c754e9a9b30"} Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.871976 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tt7cz" podStartSLOduration=2.322314302 podStartE2EDuration="15.871960038s" podCreationTimestamp="2026-03-14 09:15:58 +0000 UTC" firstStartedPulling="2026-03-14 09:15:59.749964628 +0000 UTC m=+327.062575746" lastFinishedPulling="2026-03-14 09:16:13.299610354 +0000 UTC m=+340.612221482" observedRunningTime="2026-03-14 09:16:13.868645169 +0000 UTC m=+341.181256297" watchObservedRunningTime="2026-03-14 09:16:13.871960038 +0000 UTC m=+341.184571166" Mar 14 09:16:13 crc kubenswrapper[4843]: I0314 09:16:13.938663 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-llm5h" podStartSLOduration=2.434223874 podStartE2EDuration="15.938648509s" podCreationTimestamp="2026-03-14 09:15:58 +0000 UTC" firstStartedPulling="2026-03-14 09:15:59.750950382 +0000 UTC m=+327.063561510" lastFinishedPulling="2026-03-14 09:16:13.255375007 +0000 UTC m=+340.567986145" observedRunningTime="2026-03-14 09:16:13.935837051 +0000 UTC m=+341.248448189" watchObservedRunningTime="2026-03-14 09:16:13.938648509 +0000 UTC m=+341.251259637" Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.120226 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.194212 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9d7w\" (UniqueName: \"kubernetes.io/projected/1108730f-bc5a-472b-98d5-087d6e04ad2a-kube-api-access-m9d7w\") pod \"1108730f-bc5a-472b-98d5-087d6e04ad2a\" (UID: \"1108730f-bc5a-472b-98d5-087d6e04ad2a\") " Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.199733 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1108730f-bc5a-472b-98d5-087d6e04ad2a-kube-api-access-m9d7w" (OuterVolumeSpecName: "kube-api-access-m9d7w") pod "1108730f-bc5a-472b-98d5-087d6e04ad2a" (UID: "1108730f-bc5a-472b-98d5-087d6e04ad2a"). InnerVolumeSpecName "kube-api-access-m9d7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.296055 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9d7w\" (UniqueName: \"kubernetes.io/projected/1108730f-bc5a-472b-98d5-087d6e04ad2a-kube-api-access-m9d7w\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.862544 4843 generic.go:334] "Generic (PLEG): container finished" podID="f9c9423f-f5aa-4a79-8084-f775e9a8af97" containerID="06736abea2e812a113a76e07c6a59d639353333628be3fe9f0ea4c754e9a9b30" exitCode=0 Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.862584 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx4kc" event={"ID":"f9c9423f-f5aa-4a79-8084-f775e9a8af97","Type":"ContainerDied","Data":"06736abea2e812a113a76e07c6a59d639353333628be3fe9f0ea4c754e9a9b30"} Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.864453 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557996-tplp2" event={"ID":"1108730f-bc5a-472b-98d5-087d6e04ad2a","Type":"ContainerDied","Data":"442ec4ea34ee7dee690258878539a76f0e2731d06b304f11d5d339225959202c"} Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.864469 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557996-tplp2" Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.864485 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="442ec4ea34ee7dee690258878539a76f0e2731d06b304f11d5d339225959202c" Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.868468 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfbnl" event={"ID":"166fa22e-f961-4d02-bd2a-a6e215204f4b","Type":"ContainerStarted","Data":"4a6646c369cf47d0414c08222450a64f4aa414910d6c958da8f1f6c466e63c41"} Mar 14 09:16:14 crc kubenswrapper[4843]: I0314 09:16:14.893998 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tfbnl" podStartSLOduration=13.347629624 podStartE2EDuration="14.893980802s" podCreationTimestamp="2026-03-14 09:16:00 +0000 UTC" firstStartedPulling="2026-03-14 09:16:12.832141674 +0000 UTC m=+340.144752822" lastFinishedPulling="2026-03-14 09:16:14.378492872 +0000 UTC m=+341.691104000" observedRunningTime="2026-03-14 09:16:14.892416655 +0000 UTC m=+342.205027783" watchObservedRunningTime="2026-03-14 09:16:14.893980802 +0000 UTC m=+342.206591930" Mar 14 09:16:15 crc kubenswrapper[4843]: I0314 09:16:15.874073 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx4kc" event={"ID":"f9c9423f-f5aa-4a79-8084-f775e9a8af97","Type":"ContainerStarted","Data":"fc18dccb36ff55a097e0f57916d11be48f9bcec31ca0d1fd6c280535cea69c25"} Mar 14 09:16:15 crc kubenswrapper[4843]: I0314 09:16:15.893845 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vx4kc" podStartSLOduration=13.521037862 podStartE2EDuration="15.893829311s" podCreationTimestamp="2026-03-14 09:16:00 +0000 UTC" firstStartedPulling="2026-03-14 09:16:12.842462624 +0000 UTC m=+340.155073792" lastFinishedPulling="2026-03-14 09:16:15.215254113 +0000 UTC m=+342.527865241" observedRunningTime="2026-03-14 09:16:15.891368831 +0000 UTC m=+343.203979959" watchObservedRunningTime="2026-03-14 09:16:15.893829311 +0000 UTC m=+343.206440429" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.353070 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.353457 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.425118 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.551739 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.551796 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.594725 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.930952 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-llm5h" Mar 14 09:16:18 crc kubenswrapper[4843]: I0314 09:16:18.954948 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tt7cz" Mar 14 09:16:20 crc kubenswrapper[4843]: I0314 09:16:20.778012 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:20 crc kubenswrapper[4843]: I0314 09:16:20.778803 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:20 crc kubenswrapper[4843]: I0314 09:16:20.829179 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:20 crc kubenswrapper[4843]: I0314 09:16:20.940728 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tfbnl" Mar 14 09:16:20 crc kubenswrapper[4843]: I0314 09:16:20.969528 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:20 crc kubenswrapper[4843]: I0314 09:16:20.969576 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:22 crc kubenswrapper[4843]: I0314 09:16:22.025830 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vx4kc" podUID="f9c9423f-f5aa-4a79-8084-f775e9a8af97" containerName="registry-server" probeResult="failure" output=< Mar 14 09:16:22 crc kubenswrapper[4843]: timeout: failed to connect service ":50051" within 1s Mar 14 09:16:22 crc kubenswrapper[4843]: > Mar 14 09:16:30 crc kubenswrapper[4843]: I0314 09:16:30.378073 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8qk6m" Mar 14 09:16:30 crc kubenswrapper[4843]: I0314 09:16:30.462053 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hfz6k"] Mar 14 09:16:31 crc kubenswrapper[4843]: I0314 09:16:31.004474 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:31 crc kubenswrapper[4843]: I0314 09:16:31.053172 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vx4kc" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.506206 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" podUID="cf1e453b-16e7-4413-97f2-ebf025a6f6d0" containerName="registry" containerID="cri-o://b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569" gracePeriod=30 Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.899372 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950434 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-bound-sa-token\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950536 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-trusted-ca\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950572 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-tls\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950598 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-certificates\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950833 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950862 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g77xg\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-kube-api-access-g77xg\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950880 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-ca-trust-extracted\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.950903 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-installation-pull-secrets\") pod \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\" (UID: \"cf1e453b-16e7-4413-97f2-ebf025a6f6d0\") " Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.952030 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.952252 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.956795 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-kube-api-access-g77xg" (OuterVolumeSpecName: "kube-api-access-g77xg") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "kube-api-access-g77xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.957584 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.957876 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.959627 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.966935 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 14 09:16:55 crc kubenswrapper[4843]: I0314 09:16:55.970139 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cf1e453b-16e7-4413-97f2-ebf025a6f6d0" (UID: "cf1e453b-16e7-4413-97f2-ebf025a6f6d0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052728 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g77xg\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-kube-api-access-g77xg\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052782 4843 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052801 4843 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052820 4843 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052840 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052859 4843 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.052881 4843 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cf1e453b-16e7-4413-97f2-ebf025a6f6d0-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.107443 4843 generic.go:334] "Generic (PLEG): container finished" podID="cf1e453b-16e7-4413-97f2-ebf025a6f6d0" containerID="b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569" exitCode=0 Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.107502 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" event={"ID":"cf1e453b-16e7-4413-97f2-ebf025a6f6d0","Type":"ContainerDied","Data":"b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569"} Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.107549 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" event={"ID":"cf1e453b-16e7-4413-97f2-ebf025a6f6d0","Type":"ContainerDied","Data":"21e49ab2218cbcc757b33dc9fd636d66c2d781eeceaab0adcc5b1d5740e9273c"} Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.107577 4843 scope.go:117] "RemoveContainer" containerID="b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.107506 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hfz6k" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.150346 4843 scope.go:117] "RemoveContainer" containerID="b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569" Mar 14 09:16:56 crc kubenswrapper[4843]: E0314 09:16:56.152793 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569\": container with ID starting with b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569 not found: ID does not exist" containerID="b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.152842 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569"} err="failed to get container status \"b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569\": rpc error: code = NotFound desc = could not find container \"b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569\": container with ID starting with b589bef0b23039d997c758adf382416bc381d9c9366fcc3930feea6d39f7f569 not found: ID does not exist" Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.155147 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hfz6k"] Mar 14 09:16:56 crc kubenswrapper[4843]: I0314 09:16:56.160436 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hfz6k"] Mar 14 09:16:57 crc kubenswrapper[4843]: I0314 09:16:57.346120 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf1e453b-16e7-4413-97f2-ebf025a6f6d0" path="/var/lib/kubelet/pods/cf1e453b-16e7-4413-97f2-ebf025a6f6d0/volumes" Mar 14 09:17:15 crc kubenswrapper[4843]: I0314 09:17:15.387860 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:17:15 crc kubenswrapper[4843]: I0314 09:17:15.388481 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:17:45 crc kubenswrapper[4843]: I0314 09:17:45.387342 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:17:45 crc kubenswrapper[4843]: I0314 09:17:45.387917 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.132630 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557998-pjtf2"] Mar 14 09:18:00 crc kubenswrapper[4843]: E0314 09:18:00.133374 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1108730f-bc5a-472b-98d5-087d6e04ad2a" containerName="oc" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.133389 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1108730f-bc5a-472b-98d5-087d6e04ad2a" containerName="oc" Mar 14 09:18:00 crc kubenswrapper[4843]: E0314 09:18:00.133402 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf1e453b-16e7-4413-97f2-ebf025a6f6d0" containerName="registry" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.133409 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf1e453b-16e7-4413-97f2-ebf025a6f6d0" containerName="registry" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.133532 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1108730f-bc5a-472b-98d5-087d6e04ad2a" containerName="oc" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.133546 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf1e453b-16e7-4413-97f2-ebf025a6f6d0" containerName="registry" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.134019 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.136727 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.137619 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.137937 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.141805 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557998-pjtf2"] Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.312700 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxhp8\" (UniqueName: \"kubernetes.io/projected/494e5ab2-0422-4ac9-b963-15303eb3dcea-kube-api-access-sxhp8\") pod \"auto-csr-approver-29557998-pjtf2\" (UID: \"494e5ab2-0422-4ac9-b963-15303eb3dcea\") " pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.414189 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxhp8\" (UniqueName: \"kubernetes.io/projected/494e5ab2-0422-4ac9-b963-15303eb3dcea-kube-api-access-sxhp8\") pod \"auto-csr-approver-29557998-pjtf2\" (UID: \"494e5ab2-0422-4ac9-b963-15303eb3dcea\") " pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.446992 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxhp8\" (UniqueName: \"kubernetes.io/projected/494e5ab2-0422-4ac9-b963-15303eb3dcea-kube-api-access-sxhp8\") pod \"auto-csr-approver-29557998-pjtf2\" (UID: \"494e5ab2-0422-4ac9-b963-15303eb3dcea\") " pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.452139 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.631499 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557998-pjtf2"] Mar 14 09:18:00 crc kubenswrapper[4843]: I0314 09:18:00.641261 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:18:01 crc kubenswrapper[4843]: I0314 09:18:01.482079 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" event={"ID":"494e5ab2-0422-4ac9-b963-15303eb3dcea","Type":"ContainerStarted","Data":"15aafa7d9c102417faa5daf71929da3b3514fdbae9d5aa153160787140b2e5f3"} Mar 14 09:18:02 crc kubenswrapper[4843]: I0314 09:18:02.489562 4843 generic.go:334] "Generic (PLEG): container finished" podID="494e5ab2-0422-4ac9-b963-15303eb3dcea" containerID="9a649913f4f7f5d84c93acf1d9dafd9a61e47d9b2247f5e7c364c17ed852a9e3" exitCode=0 Mar 14 09:18:02 crc kubenswrapper[4843]: I0314 09:18:02.489613 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" event={"ID":"494e5ab2-0422-4ac9-b963-15303eb3dcea","Type":"ContainerDied","Data":"9a649913f4f7f5d84c93acf1d9dafd9a61e47d9b2247f5e7c364c17ed852a9e3"} Mar 14 09:18:03 crc kubenswrapper[4843]: I0314 09:18:03.695555 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:03 crc kubenswrapper[4843]: I0314 09:18:03.869313 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxhp8\" (UniqueName: \"kubernetes.io/projected/494e5ab2-0422-4ac9-b963-15303eb3dcea-kube-api-access-sxhp8\") pod \"494e5ab2-0422-4ac9-b963-15303eb3dcea\" (UID: \"494e5ab2-0422-4ac9-b963-15303eb3dcea\") " Mar 14 09:18:03 crc kubenswrapper[4843]: I0314 09:18:03.877397 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/494e5ab2-0422-4ac9-b963-15303eb3dcea-kube-api-access-sxhp8" (OuterVolumeSpecName: "kube-api-access-sxhp8") pod "494e5ab2-0422-4ac9-b963-15303eb3dcea" (UID: "494e5ab2-0422-4ac9-b963-15303eb3dcea"). InnerVolumeSpecName "kube-api-access-sxhp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:18:03 crc kubenswrapper[4843]: I0314 09:18:03.970581 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxhp8\" (UniqueName: \"kubernetes.io/projected/494e5ab2-0422-4ac9-b963-15303eb3dcea-kube-api-access-sxhp8\") on node \"crc\" DevicePath \"\"" Mar 14 09:18:04 crc kubenswrapper[4843]: I0314 09:18:04.504926 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" event={"ID":"494e5ab2-0422-4ac9-b963-15303eb3dcea","Type":"ContainerDied","Data":"15aafa7d9c102417faa5daf71929da3b3514fdbae9d5aa153160787140b2e5f3"} Mar 14 09:18:04 crc kubenswrapper[4843]: I0314 09:18:04.505007 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15aafa7d9c102417faa5daf71929da3b3514fdbae9d5aa153160787140b2e5f3" Mar 14 09:18:04 crc kubenswrapper[4843]: I0314 09:18:04.505017 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557998-pjtf2" Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.387714 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.388198 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.388431 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.389082 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb6c919a4c034719ae1f3b02fb3d4bb9d30d21fd25a1a82acf24c004a65a3907"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.389130 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://bb6c919a4c034719ae1f3b02fb3d4bb9d30d21fd25a1a82acf24c004a65a3907" gracePeriod=600 Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.579508 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="bb6c919a4c034719ae1f3b02fb3d4bb9d30d21fd25a1a82acf24c004a65a3907" exitCode=0 Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.579575 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"bb6c919a4c034719ae1f3b02fb3d4bb9d30d21fd25a1a82acf24c004a65a3907"} Mar 14 09:18:15 crc kubenswrapper[4843]: I0314 09:18:15.579626 4843 scope.go:117] "RemoveContainer" containerID="1eec10ad2604b241d08a4efa3556d36326f8f05b555f3f2dc5588bc1ea55449f" Mar 14 09:18:16 crc kubenswrapper[4843]: I0314 09:18:16.587909 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"11847d99c55208b09be55f12335af74a9e648d402160cb65deed3a9467da14ac"} Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.145051 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558000-dj6lc"] Mar 14 09:20:00 crc kubenswrapper[4843]: E0314 09:20:00.145925 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="494e5ab2-0422-4ac9-b963-15303eb3dcea" containerName="oc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.145941 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="494e5ab2-0422-4ac9-b963-15303eb3dcea" containerName="oc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.146041 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="494e5ab2-0422-4ac9-b963-15303eb3dcea" containerName="oc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.146550 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.148797 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.149147 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.149455 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.154831 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558000-dj6lc"] Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.341292 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4dnm\" (UniqueName: \"kubernetes.io/projected/9ac4bb14-edec-4967-a71f-20e67f264a0a-kube-api-access-z4dnm\") pod \"auto-csr-approver-29558000-dj6lc\" (UID: \"9ac4bb14-edec-4967-a71f-20e67f264a0a\") " pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.442686 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4dnm\" (UniqueName: \"kubernetes.io/projected/9ac4bb14-edec-4967-a71f-20e67f264a0a-kube-api-access-z4dnm\") pod \"auto-csr-approver-29558000-dj6lc\" (UID: \"9ac4bb14-edec-4967-a71f-20e67f264a0a\") " pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.465686 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4dnm\" (UniqueName: \"kubernetes.io/projected/9ac4bb14-edec-4967-a71f-20e67f264a0a-kube-api-access-z4dnm\") pod \"auto-csr-approver-29558000-dj6lc\" (UID: \"9ac4bb14-edec-4967-a71f-20e67f264a0a\") " pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.472115 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:00 crc kubenswrapper[4843]: I0314 09:20:00.891947 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558000-dj6lc"] Mar 14 09:20:00 crc kubenswrapper[4843]: W0314 09:20:00.900471 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ac4bb14_edec_4967_a71f_20e67f264a0a.slice/crio-ba9b4b12366fdb99d00da0ea0c2b6d0005178e0834a7a562648cc4d28561347f WatchSource:0}: Error finding container ba9b4b12366fdb99d00da0ea0c2b6d0005178e0834a7a562648cc4d28561347f: Status 404 returned error can't find the container with id ba9b4b12366fdb99d00da0ea0c2b6d0005178e0834a7a562648cc4d28561347f Mar 14 09:20:01 crc kubenswrapper[4843]: I0314 09:20:01.180782 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" event={"ID":"9ac4bb14-edec-4967-a71f-20e67f264a0a","Type":"ContainerStarted","Data":"ba9b4b12366fdb99d00da0ea0c2b6d0005178e0834a7a562648cc4d28561347f"} Mar 14 09:20:02 crc kubenswrapper[4843]: I0314 09:20:02.187700 4843 generic.go:334] "Generic (PLEG): container finished" podID="9ac4bb14-edec-4967-a71f-20e67f264a0a" containerID="46385c8fe57811cadbd73b503074c929b631e8a732af4004508a58abc3c155f7" exitCode=0 Mar 14 09:20:02 crc kubenswrapper[4843]: I0314 09:20:02.187749 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" event={"ID":"9ac4bb14-edec-4967-a71f-20e67f264a0a","Type":"ContainerDied","Data":"46385c8fe57811cadbd73b503074c929b631e8a732af4004508a58abc3c155f7"} Mar 14 09:20:03 crc kubenswrapper[4843]: I0314 09:20:03.398196 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:03 crc kubenswrapper[4843]: I0314 09:20:03.580346 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4dnm\" (UniqueName: \"kubernetes.io/projected/9ac4bb14-edec-4967-a71f-20e67f264a0a-kube-api-access-z4dnm\") pod \"9ac4bb14-edec-4967-a71f-20e67f264a0a\" (UID: \"9ac4bb14-edec-4967-a71f-20e67f264a0a\") " Mar 14 09:20:03 crc kubenswrapper[4843]: I0314 09:20:03.589653 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac4bb14-edec-4967-a71f-20e67f264a0a-kube-api-access-z4dnm" (OuterVolumeSpecName: "kube-api-access-z4dnm") pod "9ac4bb14-edec-4967-a71f-20e67f264a0a" (UID: "9ac4bb14-edec-4967-a71f-20e67f264a0a"). InnerVolumeSpecName "kube-api-access-z4dnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:20:03 crc kubenswrapper[4843]: I0314 09:20:03.681831 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4dnm\" (UniqueName: \"kubernetes.io/projected/9ac4bb14-edec-4967-a71f-20e67f264a0a-kube-api-access-z4dnm\") on node \"crc\" DevicePath \"\"" Mar 14 09:20:04 crc kubenswrapper[4843]: I0314 09:20:04.200348 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" event={"ID":"9ac4bb14-edec-4967-a71f-20e67f264a0a","Type":"ContainerDied","Data":"ba9b4b12366fdb99d00da0ea0c2b6d0005178e0834a7a562648cc4d28561347f"} Mar 14 09:20:04 crc kubenswrapper[4843]: I0314 09:20:04.200659 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba9b4b12366fdb99d00da0ea0c2b6d0005178e0834a7a562648cc4d28561347f" Mar 14 09:20:04 crc kubenswrapper[4843]: I0314 09:20:04.200413 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558000-dj6lc" Mar 14 09:20:04 crc kubenswrapper[4843]: I0314 09:20:04.462687 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557994-q425x"] Mar 14 09:20:04 crc kubenswrapper[4843]: I0314 09:20:04.469329 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557994-q425x"] Mar 14 09:20:05 crc kubenswrapper[4843]: I0314 09:20:05.353682 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f190b327-cd1f-4a28-9d0b-7e1980a334a6" path="/var/lib/kubelet/pods/f190b327-cd1f-4a28-9d0b-7e1980a334a6/volumes" Mar 14 09:20:15 crc kubenswrapper[4843]: I0314 09:20:15.387090 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:20:15 crc kubenswrapper[4843]: I0314 09:20:15.387947 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:20:40 crc kubenswrapper[4843]: I0314 09:20:40.280171 4843 scope.go:117] "RemoveContainer" containerID="5bca3d22e3db02690b7de5f6b625153e094fafb851e9b5c4f149e659213683b8" Mar 14 09:20:45 crc kubenswrapper[4843]: I0314 09:20:45.387541 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:20:45 crc kubenswrapper[4843]: I0314 09:20:45.388132 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.387432 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.388055 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.388142 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.389051 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11847d99c55208b09be55f12335af74a9e648d402160cb65deed3a9467da14ac"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.389215 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://11847d99c55208b09be55f12335af74a9e648d402160cb65deed3a9467da14ac" gracePeriod=600 Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.640329 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="11847d99c55208b09be55f12335af74a9e648d402160cb65deed3a9467da14ac" exitCode=0 Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.640388 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"11847d99c55208b09be55f12335af74a9e648d402160cb65deed3a9467da14ac"} Mar 14 09:21:15 crc kubenswrapper[4843]: I0314 09:21:15.640432 4843 scope.go:117] "RemoveContainer" containerID="bb6c919a4c034719ae1f3b02fb3d4bb9d30d21fd25a1a82acf24c004a65a3907" Mar 14 09:21:16 crc kubenswrapper[4843]: I0314 09:21:16.663211 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"0419057bafaf42ee3e8fe2dcca9e3ad2340132e208a0d8b9f235557b85f11481"} Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.781617 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7"] Mar 14 09:21:34 crc kubenswrapper[4843]: E0314 09:21:34.782412 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac4bb14-edec-4967-a71f-20e67f264a0a" containerName="oc" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.782431 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac4bb14-edec-4967-a71f-20e67f264a0a" containerName="oc" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.782590 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac4bb14-edec-4967-a71f-20e67f264a0a" containerName="oc" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.783603 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.786977 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.795662 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7"] Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.866433 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.866552 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.866725 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdptm\" (UniqueName: \"kubernetes.io/projected/cb06d762-fe54-4e6f-917e-1431b5948d9a-kube-api-access-bdptm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.968118 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdptm\" (UniqueName: \"kubernetes.io/projected/cb06d762-fe54-4e6f-917e-1431b5948d9a-kube-api-access-bdptm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.968201 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.968236 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.968782 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.968845 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:34 crc kubenswrapper[4843]: I0314 09:21:34.989417 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdptm\" (UniqueName: \"kubernetes.io/projected/cb06d762-fe54-4e6f-917e-1431b5948d9a-kube-api-access-bdptm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:35 crc kubenswrapper[4843]: I0314 09:21:35.102362 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:35 crc kubenswrapper[4843]: I0314 09:21:35.300792 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7"] Mar 14 09:21:35 crc kubenswrapper[4843]: I0314 09:21:35.773707 4843 generic.go:334] "Generic (PLEG): container finished" podID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerID="11159e243db3574b17f16bb9469116205c16100d4aeea772bf6486f418677916" exitCode=0 Mar 14 09:21:35 crc kubenswrapper[4843]: I0314 09:21:35.773756 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" event={"ID":"cb06d762-fe54-4e6f-917e-1431b5948d9a","Type":"ContainerDied","Data":"11159e243db3574b17f16bb9469116205c16100d4aeea772bf6486f418677916"} Mar 14 09:21:35 crc kubenswrapper[4843]: I0314 09:21:35.773982 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" event={"ID":"cb06d762-fe54-4e6f-917e-1431b5948d9a","Type":"ContainerStarted","Data":"16a76ff8eb39068b702a211fa3eefd590497bb1cf2c31f787f94ffc38e014bad"} Mar 14 09:21:37 crc kubenswrapper[4843]: I0314 09:21:37.786850 4843 generic.go:334] "Generic (PLEG): container finished" podID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerID="f58a9c1162f791361af1d4d3ee32bff529bec57913c86745e1091c2e5efda00e" exitCode=0 Mar 14 09:21:37 crc kubenswrapper[4843]: I0314 09:21:37.786915 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" event={"ID":"cb06d762-fe54-4e6f-917e-1431b5948d9a","Type":"ContainerDied","Data":"f58a9c1162f791361af1d4d3ee32bff529bec57913c86745e1091c2e5efda00e"} Mar 14 09:21:38 crc kubenswrapper[4843]: I0314 09:21:38.796633 4843 generic.go:334] "Generic (PLEG): container finished" podID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerID="699976f96e9d2e8a938ce6a4874543da924c9e530b36705a14ad23cf07bc2dfd" exitCode=0 Mar 14 09:21:38 crc kubenswrapper[4843]: I0314 09:21:38.796714 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" event={"ID":"cb06d762-fe54-4e6f-917e-1431b5948d9a","Type":"ContainerDied","Data":"699976f96e9d2e8a938ce6a4874543da924c9e530b36705a14ad23cf07bc2dfd"} Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.100659 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.136571 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-bundle\") pod \"cb06d762-fe54-4e6f-917e-1431b5948d9a\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.136718 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdptm\" (UniqueName: \"kubernetes.io/projected/cb06d762-fe54-4e6f-917e-1431b5948d9a-kube-api-access-bdptm\") pod \"cb06d762-fe54-4e6f-917e-1431b5948d9a\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.136795 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-util\") pod \"cb06d762-fe54-4e6f-917e-1431b5948d9a\" (UID: \"cb06d762-fe54-4e6f-917e-1431b5948d9a\") " Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.139138 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-bundle" (OuterVolumeSpecName: "bundle") pod "cb06d762-fe54-4e6f-917e-1431b5948d9a" (UID: "cb06d762-fe54-4e6f-917e-1431b5948d9a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.143320 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb06d762-fe54-4e6f-917e-1431b5948d9a-kube-api-access-bdptm" (OuterVolumeSpecName: "kube-api-access-bdptm") pod "cb06d762-fe54-4e6f-917e-1431b5948d9a" (UID: "cb06d762-fe54-4e6f-917e-1431b5948d9a"). InnerVolumeSpecName "kube-api-access-bdptm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.162504 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-util" (OuterVolumeSpecName: "util") pod "cb06d762-fe54-4e6f-917e-1431b5948d9a" (UID: "cb06d762-fe54-4e6f-917e-1431b5948d9a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.237832 4843 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.237858 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdptm\" (UniqueName: \"kubernetes.io/projected/cb06d762-fe54-4e6f-917e-1431b5948d9a-kube-api-access-bdptm\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.237867 4843 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cb06d762-fe54-4e6f-917e-1431b5948d9a-util\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.810742 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" event={"ID":"cb06d762-fe54-4e6f-917e-1431b5948d9a","Type":"ContainerDied","Data":"16a76ff8eb39068b702a211fa3eefd590497bb1cf2c31f787f94ffc38e014bad"} Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.810783 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16a76ff8eb39068b702a211fa3eefd590497bb1cf2c31f787f94ffc38e014bad" Mar 14 09:21:40 crc kubenswrapper[4843]: I0314 09:21:40.810791 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7" Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.726437 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vfpmm"] Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.727413 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-controller" containerID="cri-o://be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.727991 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="sbdb" containerID="cri-o://749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.728051 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="nbdb" containerID="cri-o://4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.728092 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="northd" containerID="cri-o://d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.728156 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.728203 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-node" containerID="cri-o://72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.728242 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-acl-logging" containerID="cri-o://8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.768594 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovnkube-controller" containerID="cri-o://9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" gracePeriod=30 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.836102 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sd9qb_32cf92ff-937a-41d1-b27a-127bd550a1eb/kube-multus/0.log" Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.836584 4843 generic.go:334] "Generic (PLEG): container finished" podID="32cf92ff-937a-41d1-b27a-127bd550a1eb" containerID="e93284b6b38fee95cbefc6bd220dc1328863323ce552a6eb6bed85a982184a8c" exitCode=2 Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.836627 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sd9qb" event={"ID":"32cf92ff-937a-41d1-b27a-127bd550a1eb","Type":"ContainerDied","Data":"e93284b6b38fee95cbefc6bd220dc1328863323ce552a6eb6bed85a982184a8c"} Mar 14 09:21:45 crc kubenswrapper[4843]: I0314 09:21:45.837305 4843 scope.go:117] "RemoveContainer" containerID="e93284b6b38fee95cbefc6bd220dc1328863323ce552a6eb6bed85a982184a8c" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.077575 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vfpmm_986f5a7d-244f-4068-bd4d-949b7afc1e16/ovn-acl-logging/0.log" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.077952 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vfpmm_986f5a7d-244f-4068-bd4d-949b7afc1e16/ovn-controller/0.log" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.078314 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113450 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113500 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqwqh\" (UniqueName: \"kubernetes.io/projected/986f5a7d-244f-4068-bd4d-949b7afc1e16-kube-api-access-gqwqh\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113525 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-config\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113544 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-netd\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113572 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-slash\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113589 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-node-log\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113604 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-env-overrides\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113622 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-var-lib-openvswitch\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113637 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-openvswitch\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113665 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-log-socket\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113682 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-script-lib\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113707 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovn-node-metrics-cert\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113734 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-netns\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113759 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-ovn-kubernetes\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113773 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-etc-openvswitch\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113786 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-bin\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113802 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-systemd-units\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113815 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-kubelet\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113828 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-systemd\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.113843 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-ovn\") pod \"986f5a7d-244f-4068-bd4d-949b7afc1e16\" (UID: \"986f5a7d-244f-4068-bd4d-949b7afc1e16\") " Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.114035 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.114065 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.114771 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.114881 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.114953 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.114993 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115006 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115019 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-log-socket" (OuterVolumeSpecName: "log-socket") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115036 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-slash" (OuterVolumeSpecName: "host-slash") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115048 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115063 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115075 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115092 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115305 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115319 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115334 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.115344 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-node-log" (OuterVolumeSpecName: "node-log") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.126119 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/986f5a7d-244f-4068-bd4d-949b7afc1e16-kube-api-access-gqwqh" (OuterVolumeSpecName: "kube-api-access-gqwqh") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "kube-api-access-gqwqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.126560 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.133523 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "986f5a7d-244f-4068-bd4d-949b7afc1e16" (UID: "986f5a7d-244f-4068-bd4d-949b7afc1e16"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.159282 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rqzbl"] Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.159722 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="pull" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.159797 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="pull" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.159848 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="sbdb" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.159900 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="sbdb" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.159951 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kubecfg-setup" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.159996 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kubecfg-setup" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160040 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="northd" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160082 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="northd" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160129 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-controller" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160172 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-controller" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160224 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="util" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160281 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="util" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160328 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="nbdb" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160378 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="nbdb" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160430 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-ovn-metrics" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160477 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-ovn-metrics" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160528 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovnkube-controller" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160578 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovnkube-controller" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160625 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-acl-logging" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160672 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-acl-logging" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160719 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-node" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160761 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-node" Mar 14 09:21:46 crc kubenswrapper[4843]: E0314 09:21:46.160820 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="extract" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.160877 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="extract" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161016 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="northd" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161077 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovnkube-controller" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161147 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-acl-logging" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161307 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="sbdb" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161390 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb06d762-fe54-4e6f-917e-1431b5948d9a" containerName="extract" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161461 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="ovn-controller" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161538 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-ovn-metrics" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161586 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="nbdb" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.161639 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerName="kube-rbac-proxy-node" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.163334 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214611 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-node-log\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214655 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214676 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovnkube-script-lib\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214699 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfwh4\" (UniqueName: \"kubernetes.io/projected/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-kube-api-access-hfwh4\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214714 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-log-socket\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214736 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-cni-bin\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214752 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214784 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-var-lib-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214812 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-env-overrides\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214842 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-ovn\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214863 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-systemd\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214881 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-kubelet\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214902 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-run-netns\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214924 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-slash\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214955 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-cni-netd\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214980 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-systemd-units\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.214998 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovnkube-config\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215021 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovn-node-metrics-cert\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215040 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-etc-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215058 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-run-ovn-kubernetes\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215104 4843 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215127 4843 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215164 4843 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215177 4843 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215188 4843 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215199 4843 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215206 4843 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215214 4843 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215222 4843 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215232 4843 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215241 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqwqh\" (UniqueName: \"kubernetes.io/projected/986f5a7d-244f-4068-bd4d-949b7afc1e16-kube-api-access-gqwqh\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215255 4843 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215265 4843 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215294 4843 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-host-slash\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215305 4843 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-node-log\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215314 4843 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215325 4843 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215335 4843 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215344 4843 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/986f5a7d-244f-4068-bd4d-949b7afc1e16-log-socket\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.215355 4843 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/986f5a7d-244f-4068-bd4d-949b7afc1e16-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316137 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-systemd-units\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316177 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovnkube-config\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316202 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovn-node-metrics-cert\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316217 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-etc-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316233 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-run-ovn-kubernetes\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316228 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-systemd-units\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316253 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-node-log\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316323 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316346 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-node-log\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316354 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovnkube-script-lib\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316376 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-etc-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316383 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfwh4\" (UniqueName: \"kubernetes.io/projected/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-kube-api-access-hfwh4\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316397 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-run-ovn-kubernetes\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316402 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-log-socket\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316424 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-log-socket\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316432 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-cni-bin\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316449 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316453 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316472 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316477 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-var-lib-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316490 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-cni-bin\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316498 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-env-overrides\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316520 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-ovn\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316536 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-systemd\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316553 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-kubelet\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316569 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-run-netns\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316589 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-slash\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316606 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-cni-netd\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.316649 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-cni-netd\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317118 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovnkube-config\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317155 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovnkube-script-lib\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317156 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-ovn\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317197 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-run-systemd\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317217 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-kubelet\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317236 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-run-netns\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317256 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-host-slash\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317298 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-var-lib-openvswitch\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.317591 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-env-overrides\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.319854 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-ovn-node-metrics-cert\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.340364 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfwh4\" (UniqueName: \"kubernetes.io/projected/bc4715f8-9fa2-474e-b1cd-94ceeeff9505-kube-api-access-hfwh4\") pod \"ovnkube-node-rqzbl\" (UID: \"bc4715f8-9fa2-474e-b1cd-94ceeeff9505\") " pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.475955 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:46 crc kubenswrapper[4843]: W0314 09:21:46.492642 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc4715f8_9fa2_474e_b1cd_94ceeeff9505.slice/crio-b5f262abce9023595a14678f63e435edc2cbc7ccba2cb5bbf07000ff65ff80cc WatchSource:0}: Error finding container b5f262abce9023595a14678f63e435edc2cbc7ccba2cb5bbf07000ff65ff80cc: Status 404 returned error can't find the container with id b5f262abce9023595a14678f63e435edc2cbc7ccba2cb5bbf07000ff65ff80cc Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.844894 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vfpmm_986f5a7d-244f-4068-bd4d-949b7afc1e16/ovn-acl-logging/0.log" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.846381 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vfpmm_986f5a7d-244f-4068-bd4d-949b7afc1e16/ovn-controller/0.log" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.846825 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.846919 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.846992 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.846966 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847056 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847287 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847373 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847449 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" exitCode=143 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847512 4843 generic.go:334] "Generic (PLEG): container finished" podID="986f5a7d-244f-4068-bd4d-949b7afc1e16" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" exitCode=143 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.846872 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847713 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847896 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847977 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848358 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848665 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.847911 4843 scope.go:117] "RemoveContainer" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848811 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848918 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848936 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848954 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848977 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848987 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848993 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.848998 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849003 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849009 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849014 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849019 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849024 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849031 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849040 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849048 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849054 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849062 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849069 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849074 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849079 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849085 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849090 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849098 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vfpmm" event={"ID":"986f5a7d-244f-4068-bd4d-949b7afc1e16","Type":"ContainerDied","Data":"63d446a30e03b814adb1e4107ae699b45215b3e3da41ad7fb3a55c457707f1a6"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849114 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849120 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849126 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849131 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849136 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849141 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849146 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849151 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.849155 4843 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.850659 4843 generic.go:334] "Generic (PLEG): container finished" podID="bc4715f8-9fa2-474e-b1cd-94ceeeff9505" containerID="57412c05900d9fa4ec92a2a1a504990fc0a7bb421f32c7301408dc20679d8c63" exitCode=0 Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.850700 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerDied","Data":"57412c05900d9fa4ec92a2a1a504990fc0a7bb421f32c7301408dc20679d8c63"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.850715 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"b5f262abce9023595a14678f63e435edc2cbc7ccba2cb5bbf07000ff65ff80cc"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.853440 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sd9qb_32cf92ff-937a-41d1-b27a-127bd550a1eb/kube-multus/0.log" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.853494 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sd9qb" event={"ID":"32cf92ff-937a-41d1-b27a-127bd550a1eb","Type":"ContainerStarted","Data":"c6ff1af496e26911c38dc498cad29cfe19f7d773933d8f699d6a12f387637b31"} Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.918421 4843 scope.go:117] "RemoveContainer" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.947759 4843 scope.go:117] "RemoveContainer" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.958646 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vfpmm"] Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.966141 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vfpmm"] Mar 14 09:21:46 crc kubenswrapper[4843]: I0314 09:21:46.984343 4843 scope.go:117] "RemoveContainer" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.004396 4843 scope.go:117] "RemoveContainer" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.026226 4843 scope.go:117] "RemoveContainer" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.047954 4843 scope.go:117] "RemoveContainer" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.065706 4843 scope.go:117] "RemoveContainer" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.094946 4843 scope.go:117] "RemoveContainer" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.133318 4843 scope.go:117] "RemoveContainer" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.133684 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": container with ID starting with 9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b not found: ID does not exist" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.133722 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} err="failed to get container status \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": rpc error: code = NotFound desc = could not find container \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": container with ID starting with 9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.133746 4843 scope.go:117] "RemoveContainer" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.134609 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": container with ID starting with 749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623 not found: ID does not exist" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.134653 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} err="failed to get container status \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": rpc error: code = NotFound desc = could not find container \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": container with ID starting with 749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.134682 4843 scope.go:117] "RemoveContainer" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.134990 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": container with ID starting with 4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655 not found: ID does not exist" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.135010 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} err="failed to get container status \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": rpc error: code = NotFound desc = could not find container \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": container with ID starting with 4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.135024 4843 scope.go:117] "RemoveContainer" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.138759 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": container with ID starting with d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438 not found: ID does not exist" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.138795 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} err="failed to get container status \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": rpc error: code = NotFound desc = could not find container \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": container with ID starting with d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.138819 4843 scope.go:117] "RemoveContainer" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.145862 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": container with ID starting with d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168 not found: ID does not exist" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.145922 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} err="failed to get container status \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": rpc error: code = NotFound desc = could not find container \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": container with ID starting with d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.145957 4843 scope.go:117] "RemoveContainer" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.146695 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": container with ID starting with 72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e not found: ID does not exist" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.146719 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} err="failed to get container status \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": rpc error: code = NotFound desc = could not find container \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": container with ID starting with 72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.146734 4843 scope.go:117] "RemoveContainer" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.149139 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": container with ID starting with 8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc not found: ID does not exist" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.149167 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} err="failed to get container status \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": rpc error: code = NotFound desc = could not find container \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": container with ID starting with 8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.149182 4843 scope.go:117] "RemoveContainer" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.152938 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": container with ID starting with be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32 not found: ID does not exist" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.152983 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} err="failed to get container status \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": rpc error: code = NotFound desc = could not find container \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": container with ID starting with be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.153009 4843 scope.go:117] "RemoveContainer" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" Mar 14 09:21:47 crc kubenswrapper[4843]: E0314 09:21:47.156750 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": container with ID starting with 5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48 not found: ID does not exist" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.156820 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} err="failed to get container status \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": rpc error: code = NotFound desc = could not find container \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": container with ID starting with 5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.156853 4843 scope.go:117] "RemoveContainer" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.161464 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} err="failed to get container status \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": rpc error: code = NotFound desc = could not find container \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": container with ID starting with 9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.161509 4843 scope.go:117] "RemoveContainer" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.162035 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} err="failed to get container status \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": rpc error: code = NotFound desc = could not find container \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": container with ID starting with 749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.162072 4843 scope.go:117] "RemoveContainer" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.162373 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} err="failed to get container status \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": rpc error: code = NotFound desc = could not find container \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": container with ID starting with 4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.162402 4843 scope.go:117] "RemoveContainer" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.162777 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} err="failed to get container status \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": rpc error: code = NotFound desc = could not find container \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": container with ID starting with d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.162813 4843 scope.go:117] "RemoveContainer" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.163041 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} err="failed to get container status \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": rpc error: code = NotFound desc = could not find container \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": container with ID starting with d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.163071 4843 scope.go:117] "RemoveContainer" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.164909 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} err="failed to get container status \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": rpc error: code = NotFound desc = could not find container \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": container with ID starting with 72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.164937 4843 scope.go:117] "RemoveContainer" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.165203 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} err="failed to get container status \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": rpc error: code = NotFound desc = could not find container \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": container with ID starting with 8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.165233 4843 scope.go:117] "RemoveContainer" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.166662 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} err="failed to get container status \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": rpc error: code = NotFound desc = could not find container \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": container with ID starting with be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.166683 4843 scope.go:117] "RemoveContainer" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.168433 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} err="failed to get container status \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": rpc error: code = NotFound desc = could not find container \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": container with ID starting with 5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.168471 4843 scope.go:117] "RemoveContainer" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.171826 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} err="failed to get container status \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": rpc error: code = NotFound desc = could not find container \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": container with ID starting with 9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.171866 4843 scope.go:117] "RemoveContainer" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.172912 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} err="failed to get container status \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": rpc error: code = NotFound desc = could not find container \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": container with ID starting with 749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.172967 4843 scope.go:117] "RemoveContainer" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.173500 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} err="failed to get container status \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": rpc error: code = NotFound desc = could not find container \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": container with ID starting with 4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.173556 4843 scope.go:117] "RemoveContainer" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.176825 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} err="failed to get container status \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": rpc error: code = NotFound desc = could not find container \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": container with ID starting with d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.176860 4843 scope.go:117] "RemoveContainer" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.178848 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} err="failed to get container status \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": rpc error: code = NotFound desc = could not find container \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": container with ID starting with d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.178871 4843 scope.go:117] "RemoveContainer" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.182992 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} err="failed to get container status \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": rpc error: code = NotFound desc = could not find container \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": container with ID starting with 72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.183022 4843 scope.go:117] "RemoveContainer" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.183512 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} err="failed to get container status \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": rpc error: code = NotFound desc = could not find container \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": container with ID starting with 8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.183564 4843 scope.go:117] "RemoveContainer" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.184498 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} err="failed to get container status \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": rpc error: code = NotFound desc = could not find container \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": container with ID starting with be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.184525 4843 scope.go:117] "RemoveContainer" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.185545 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} err="failed to get container status \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": rpc error: code = NotFound desc = could not find container \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": container with ID starting with 5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.185573 4843 scope.go:117] "RemoveContainer" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.195559 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} err="failed to get container status \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": rpc error: code = NotFound desc = could not find container \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": container with ID starting with 9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.195651 4843 scope.go:117] "RemoveContainer" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.196803 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} err="failed to get container status \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": rpc error: code = NotFound desc = could not find container \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": container with ID starting with 749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.196835 4843 scope.go:117] "RemoveContainer" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.197213 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} err="failed to get container status \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": rpc error: code = NotFound desc = could not find container \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": container with ID starting with 4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.197239 4843 scope.go:117] "RemoveContainer" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.197634 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} err="failed to get container status \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": rpc error: code = NotFound desc = could not find container \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": container with ID starting with d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.197656 4843 scope.go:117] "RemoveContainer" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.197899 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} err="failed to get container status \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": rpc error: code = NotFound desc = could not find container \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": container with ID starting with d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.197919 4843 scope.go:117] "RemoveContainer" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.198233 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} err="failed to get container status \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": rpc error: code = NotFound desc = could not find container \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": container with ID starting with 72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.198251 4843 scope.go:117] "RemoveContainer" containerID="8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.198542 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc"} err="failed to get container status \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": rpc error: code = NotFound desc = could not find container \"8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc\": container with ID starting with 8f3c063f3fd3fd9ce8c663bb97b6e4b1bc0dc88ed52434bc2e873355fadb0cdc not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.198562 4843 scope.go:117] "RemoveContainer" containerID="be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.198829 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32"} err="failed to get container status \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": rpc error: code = NotFound desc = could not find container \"be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32\": container with ID starting with be78152270e9aeab33a59e9630e48988a93f60d83e08d027f1df85706947df32 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.198850 4843 scope.go:117] "RemoveContainer" containerID="5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.202357 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48"} err="failed to get container status \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": rpc error: code = NotFound desc = could not find container \"5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48\": container with ID starting with 5c80bd8d7a7e9fe186e9f92f049f7420e4a6116d21dc2a4975bca94d96657b48 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.202382 4843 scope.go:117] "RemoveContainer" containerID="9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.202742 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b"} err="failed to get container status \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": rpc error: code = NotFound desc = could not find container \"9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b\": container with ID starting with 9dfc7a955b690dab44e7b038de6e9a4c82c6686e406dec9de2607e1dd2c7f15b not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.202764 4843 scope.go:117] "RemoveContainer" containerID="749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.203033 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623"} err="failed to get container status \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": rpc error: code = NotFound desc = could not find container \"749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623\": container with ID starting with 749fa10aed5ec8ec65589262092cd6e246a7b34f26cb429c9e75fd2d2f280623 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.203055 4843 scope.go:117] "RemoveContainer" containerID="4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.203325 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655"} err="failed to get container status \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": rpc error: code = NotFound desc = could not find container \"4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655\": container with ID starting with 4a3cf03a7cdc7f6461c25718e464cd2534e55d684df4ecf6ce725e7df3536655 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.203345 4843 scope.go:117] "RemoveContainer" containerID="d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.204784 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438"} err="failed to get container status \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": rpc error: code = NotFound desc = could not find container \"d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438\": container with ID starting with d0f2d207288afc27657ca09155556f5dd328f789d04b28ad412c476647230438 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.204810 4843 scope.go:117] "RemoveContainer" containerID="d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.209383 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168"} err="failed to get container status \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": rpc error: code = NotFound desc = could not find container \"d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168\": container with ID starting with d62c8ea81e796a17ced6ad9feb9d092c574e1979b503ad405156bcd163c3b168 not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.209412 4843 scope.go:117] "RemoveContainer" containerID="72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.211253 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e"} err="failed to get container status \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": rpc error: code = NotFound desc = could not find container \"72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e\": container with ID starting with 72040d23e79f988be4365327f832932f27ddd350e8ad23bc96a66843a29b5e4e not found: ID does not exist" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.346937 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="986f5a7d-244f-4068-bd4d-949b7afc1e16" path="/var/lib/kubelet/pods/986f5a7d-244f-4068-bd4d-949b7afc1e16/volumes" Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.863564 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"bab9c620cae2285e3721ddb4aab35f1c9164ba1bd53975d8ff90621345f18233"} Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.864640 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"8c7287a032b7fbc27f9ab7039a0b2ff7b90dee78107c0eaf7b5adc2f8520c42d"} Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.864710 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"72f8bb627d20dfe3910b1dee7496bdd1068cda65f30c30d8f6432c20c318bd0b"} Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.864762 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"6c39a1a2b6030ff8933bf3519ac074192a81b3645e355b731fd60455989ef9bc"} Mar 14 09:21:47 crc kubenswrapper[4843]: I0314 09:21:47.864820 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"67b603274d35a31636940a1b2b3dcf5dbb6d74eec596f1d6bf2649af9fc03262"} Mar 14 09:21:48 crc kubenswrapper[4843]: I0314 09:21:48.873151 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"a9464ba5c8cc71ec3c6745a11db2cf4f56ea4d22b0ad3356e7d86125b1f20269"} Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.070404 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5"] Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.071579 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.082076 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.082549 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.082919 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-jtvs7" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.126661 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk"] Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.127326 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.129014 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-ng4xw" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.129541 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.139542 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4"] Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.140393 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.212039 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d01a80d-2445-472e-ab28-81703fb66a27-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk\" (UID: \"7d01a80d-2445-472e-ab28-81703fb66a27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.212409 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d01a80d-2445-472e-ab28-81703fb66a27-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk\" (UID: \"7d01a80d-2445-472e-ab28-81703fb66a27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.212435 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl6n6\" (UniqueName: \"kubernetes.io/projected/7f706824-088d-4d70-830e-4844447b900d-kube-api-access-zl6n6\") pod \"obo-prometheus-operator-68bc856cb9-7vsx5\" (UID: \"7f706824-088d-4d70-830e-4844447b900d\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.212455 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/35d96605-ef77-45e3-b12c-26c192a2dd50-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4\" (UID: \"35d96605-ef77-45e3-b12c-26c192a2dd50\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.212572 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/35d96605-ef77-45e3-b12c-26c192a2dd50-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4\" (UID: \"35d96605-ef77-45e3-b12c-26c192a2dd50\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.313710 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/35d96605-ef77-45e3-b12c-26c192a2dd50-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4\" (UID: \"35d96605-ef77-45e3-b12c-26c192a2dd50\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.313779 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d01a80d-2445-472e-ab28-81703fb66a27-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk\" (UID: \"7d01a80d-2445-472e-ab28-81703fb66a27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.313805 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d01a80d-2445-472e-ab28-81703fb66a27-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk\" (UID: \"7d01a80d-2445-472e-ab28-81703fb66a27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.313887 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl6n6\" (UniqueName: \"kubernetes.io/projected/7f706824-088d-4d70-830e-4844447b900d-kube-api-access-zl6n6\") pod \"obo-prometheus-operator-68bc856cb9-7vsx5\" (UID: \"7f706824-088d-4d70-830e-4844447b900d\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.313913 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/35d96605-ef77-45e3-b12c-26c192a2dd50-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4\" (UID: \"35d96605-ef77-45e3-b12c-26c192a2dd50\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.322913 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d01a80d-2445-472e-ab28-81703fb66a27-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk\" (UID: \"7d01a80d-2445-472e-ab28-81703fb66a27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.324073 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/35d96605-ef77-45e3-b12c-26c192a2dd50-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4\" (UID: \"35d96605-ef77-45e3-b12c-26c192a2dd50\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.328702 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-7l9bs"] Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.329413 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.330740 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d01a80d-2445-472e-ab28-81703fb66a27-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk\" (UID: \"7d01a80d-2445-472e-ab28-81703fb66a27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.333363 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.333618 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-b65d6" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.338943 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/35d96605-ef77-45e3-b12c-26c192a2dd50-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4\" (UID: \"35d96605-ef77-45e3-b12c-26c192a2dd50\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.354028 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl6n6\" (UniqueName: \"kubernetes.io/projected/7f706824-088d-4d70-830e-4844447b900d-kube-api-access-zl6n6\") pod \"obo-prometheus-operator-68bc856cb9-7vsx5\" (UID: \"7f706824-088d-4d70-830e-4844447b900d\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.405446 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.415920 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d8bb5f31-a201-4fa9-a297-f294c2aefe07-observability-operator-tls\") pod \"observability-operator-59bdc8b94-7l9bs\" (UID: \"d8bb5f31-a201-4fa9-a297-f294c2aefe07\") " pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.415992 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw9fx\" (UniqueName: \"kubernetes.io/projected/d8bb5f31-a201-4fa9-a297-f294c2aefe07-kube-api-access-zw9fx\") pod \"observability-operator-59bdc8b94-7l9bs\" (UID: \"d8bb5f31-a201-4fa9-a297-f294c2aefe07\") " pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.434988 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(9d1e4c88e2df6f9cd56a93d38699e28ecf74027f7e6de2c18f9c1ca7048702b7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.435060 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(9d1e4c88e2df6f9cd56a93d38699e28ecf74027f7e6de2c18f9c1ca7048702b7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.435081 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(9d1e4c88e2df6f9cd56a93d38699e28ecf74027f7e6de2c18f9c1ca7048702b7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.435129 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators(7f706824-088d-4d70-830e-4844447b900d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators(7f706824-088d-4d70-830e-4844447b900d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(9d1e4c88e2df6f9cd56a93d38699e28ecf74027f7e6de2c18f9c1ca7048702b7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" podUID="7f706824-088d-4d70-830e-4844447b900d" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.442637 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.459101 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.463704 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(7a1e01b8df02f8f379101916e20395a4b338834a11901029598cee4b6f5f372d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.463769 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(7a1e01b8df02f8f379101916e20395a4b338834a11901029598cee4b6f5f372d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.463792 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(7a1e01b8df02f8f379101916e20395a4b338834a11901029598cee4b6f5f372d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.463881 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators(7d01a80d-2445-472e-ab28-81703fb66a27)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators(7d01a80d-2445-472e-ab28-81703fb66a27)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(7a1e01b8df02f8f379101916e20395a4b338834a11901029598cee4b6f5f372d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" podUID="7d01a80d-2445-472e-ab28-81703fb66a27" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.493395 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3af9b22ac2a0c6ed018f2fcad275a2f0b1e1124190d9fd68f5c7fc951b8c295b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.493516 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3af9b22ac2a0c6ed018f2fcad275a2f0b1e1124190d9fd68f5c7fc951b8c295b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.493546 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3af9b22ac2a0c6ed018f2fcad275a2f0b1e1124190d9fd68f5c7fc951b8c295b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.493621 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators(35d96605-ef77-45e3-b12c-26c192a2dd50)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators(35d96605-ef77-45e3-b12c-26c192a2dd50)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3af9b22ac2a0c6ed018f2fcad275a2f0b1e1124190d9fd68f5c7fc951b8c295b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" podUID="35d96605-ef77-45e3-b12c-26c192a2dd50" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.502448 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-8glvj"] Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.503287 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.508876 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-vx66q" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.517214 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d8bb5f31-a201-4fa9-a297-f294c2aefe07-observability-operator-tls\") pod \"observability-operator-59bdc8b94-7l9bs\" (UID: \"d8bb5f31-a201-4fa9-a297-f294c2aefe07\") " pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.517375 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw9fx\" (UniqueName: \"kubernetes.io/projected/d8bb5f31-a201-4fa9-a297-f294c2aefe07-kube-api-access-zw9fx\") pod \"observability-operator-59bdc8b94-7l9bs\" (UID: \"d8bb5f31-a201-4fa9-a297-f294c2aefe07\") " pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.523192 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d8bb5f31-a201-4fa9-a297-f294c2aefe07-observability-operator-tls\") pod \"observability-operator-59bdc8b94-7l9bs\" (UID: \"d8bb5f31-a201-4fa9-a297-f294c2aefe07\") " pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.558317 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw9fx\" (UniqueName: \"kubernetes.io/projected/d8bb5f31-a201-4fa9-a297-f294c2aefe07-kube-api-access-zw9fx\") pod \"observability-operator-59bdc8b94-7l9bs\" (UID: \"d8bb5f31-a201-4fa9-a297-f294c2aefe07\") " pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.618407 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl8ts\" (UniqueName: \"kubernetes.io/projected/5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c-kube-api-access-pl8ts\") pod \"perses-operator-5bf474d74f-8glvj\" (UID: \"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c\") " pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.618560 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-8glvj\" (UID: \"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c\") " pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.715347 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.720013 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-8glvj\" (UID: \"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c\") " pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.720109 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl8ts\" (UniqueName: \"kubernetes.io/projected/5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c-kube-api-access-pl8ts\") pod \"perses-operator-5bf474d74f-8glvj\" (UID: \"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c\") " pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.721612 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-8glvj\" (UID: \"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c\") " pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.742736 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl8ts\" (UniqueName: \"kubernetes.io/projected/5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c-kube-api-access-pl8ts\") pod \"perses-operator-5bf474d74f-8glvj\" (UID: \"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c\") " pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.757260 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(9f3c39702d549cbc4ba25568cde0aa094b4c7da2fbf2ad04215d2b5404f89612): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.757340 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(9f3c39702d549cbc4ba25568cde0aa094b4c7da2fbf2ad04215d2b5404f89612): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.757363 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(9f3c39702d549cbc4ba25568cde0aa094b4c7da2fbf2ad04215d2b5404f89612): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.757412 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-7l9bs_openshift-operators(d8bb5f31-a201-4fa9-a297-f294c2aefe07)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-7l9bs_openshift-operators(d8bb5f31-a201-4fa9-a297-f294c2aefe07)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(9f3c39702d549cbc4ba25568cde0aa094b4c7da2fbf2ad04215d2b5404f89612): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" podUID="d8bb5f31-a201-4fa9-a297-f294c2aefe07" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.818062 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.845849 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(da25642c8d4c61c8d990f74b499881f11c29365c35c0272f95ee7b21c339e9e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.845913 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(da25642c8d4c61c8d990f74b499881f11c29365c35c0272f95ee7b21c339e9e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.845941 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(da25642c8d4c61c8d990f74b499881f11c29365c35c0272f95ee7b21c339e9e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:50 crc kubenswrapper[4843]: E0314 09:21:50.845990 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-8glvj_openshift-operators(5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-8glvj_openshift-operators(5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(da25642c8d4c61c8d990f74b499881f11c29365c35c0272f95ee7b21c339e9e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" podUID="5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c" Mar 14 09:21:50 crc kubenswrapper[4843]: I0314 09:21:50.888769 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"9a2cca6e6701bfba48e485434385bf195d4821b17e3da987996ecb1b6494a834"} Mar 14 09:21:53 crc kubenswrapper[4843]: I0314 09:21:53.907595 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" event={"ID":"bc4715f8-9fa2-474e-b1cd-94ceeeff9505","Type":"ContainerStarted","Data":"288e63bdbf2b721b4a59a9231292bf933ec463a4cd54451b1eaafac54b2ac55d"} Mar 14 09:21:53 crc kubenswrapper[4843]: I0314 09:21:53.908104 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:53 crc kubenswrapper[4843]: I0314 09:21:53.958581 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:53 crc kubenswrapper[4843]: I0314 09:21:53.980799 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" podStartSLOduration=7.980776529 podStartE2EDuration="7.980776529s" podCreationTimestamp="2026-03-14 09:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:21:53.977201539 +0000 UTC m=+681.289812687" watchObservedRunningTime="2026-03-14 09:21:53.980776529 +0000 UTC m=+681.293387657" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.584712 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk"] Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.584862 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.585378 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.589826 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5"] Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.589960 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.590455 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.600577 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-7l9bs"] Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.600707 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.601116 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.618764 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-8glvj"] Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.618895 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.619346 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.658460 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(e9691bff12bc55839e934f643e8819c44476986d3b985309eda03f82a2e77934): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.659055 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(e9691bff12bc55839e934f643e8819c44476986d3b985309eda03f82a2e77934): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.659168 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(e9691bff12bc55839e934f643e8819c44476986d3b985309eda03f82a2e77934): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.659270 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators(7d01a80d-2445-472e-ab28-81703fb66a27)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators(7d01a80d-2445-472e-ab28-81703fb66a27)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_openshift-operators_7d01a80d-2445-472e-ab28-81703fb66a27_0(e9691bff12bc55839e934f643e8819c44476986d3b985309eda03f82a2e77934): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" podUID="7d01a80d-2445-472e-ab28-81703fb66a27" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.686087 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(8e79b1b0d3be918e73e81ef24bf6032408beabe489b71f3f987f70452cf54540): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.686150 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(8e79b1b0d3be918e73e81ef24bf6032408beabe489b71f3f987f70452cf54540): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.686175 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(8e79b1b0d3be918e73e81ef24bf6032408beabe489b71f3f987f70452cf54540): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.686223 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators(7f706824-088d-4d70-830e-4844447b900d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators(7f706824-088d-4d70-830e-4844447b900d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-7vsx5_openshift-operators_7f706824-088d-4d70-830e-4844447b900d_0(8e79b1b0d3be918e73e81ef24bf6032408beabe489b71f3f987f70452cf54540): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" podUID="7f706824-088d-4d70-830e-4844447b900d" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.699454 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4"] Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.699577 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.699968 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.715246 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(fcbc051d121c18dc1a8fee8235ac130658dcbc6baa01339e7538c367900b6402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.715330 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(fcbc051d121c18dc1a8fee8235ac130658dcbc6baa01339e7538c367900b6402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.715362 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(fcbc051d121c18dc1a8fee8235ac130658dcbc6baa01339e7538c367900b6402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.715416 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-8glvj_openshift-operators(5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-8glvj_openshift-operators(5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-8glvj_openshift-operators_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c_0(fcbc051d121c18dc1a8fee8235ac130658dcbc6baa01339e7538c367900b6402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" podUID="5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.727445 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(3fb283608c72dcbdcdd6e28bc6ab0ca8672610d7d16ea267b3c71db6f4ffe997): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.727524 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(3fb283608c72dcbdcdd6e28bc6ab0ca8672610d7d16ea267b3c71db6f4ffe997): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.727599 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(3fb283608c72dcbdcdd6e28bc6ab0ca8672610d7d16ea267b3c71db6f4ffe997): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.727673 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-7l9bs_openshift-operators(d8bb5f31-a201-4fa9-a297-f294c2aefe07)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-7l9bs_openshift-operators(d8bb5f31-a201-4fa9-a297-f294c2aefe07)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-7l9bs_openshift-operators_d8bb5f31-a201-4fa9-a297-f294c2aefe07_0(3fb283608c72dcbdcdd6e28bc6ab0ca8672610d7d16ea267b3c71db6f4ffe997): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" podUID="d8bb5f31-a201-4fa9-a297-f294c2aefe07" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.778233 4843 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3b1af36008d299db01cde348636630ef1b8432b02325bb07d38f448405ada4c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.778420 4843 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3b1af36008d299db01cde348636630ef1b8432b02325bb07d38f448405ada4c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.778516 4843 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3b1af36008d299db01cde348636630ef1b8432b02325bb07d38f448405ada4c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:21:54 crc kubenswrapper[4843]: E0314 09:21:54.778637 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators(35d96605-ef77-45e3-b12c-26c192a2dd50)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators(35d96605-ef77-45e3-b12c-26c192a2dd50)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_openshift-operators_35d96605-ef77-45e3-b12c-26c192a2dd50_0(3b1af36008d299db01cde348636630ef1b8432b02325bb07d38f448405ada4c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" podUID="35d96605-ef77-45e3-b12c-26c192a2dd50" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.913029 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.913662 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:21:54 crc kubenswrapper[4843]: I0314 09:21:54.993237 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.119999 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558002-w844d"] Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.121101 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.122888 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.123453 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.124410 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.132656 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558002-w844d"] Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.247189 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrmbl\" (UniqueName: \"kubernetes.io/projected/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8-kube-api-access-hrmbl\") pod \"auto-csr-approver-29558002-w844d\" (UID: \"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8\") " pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.348006 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrmbl\" (UniqueName: \"kubernetes.io/projected/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8-kube-api-access-hrmbl\") pod \"auto-csr-approver-29558002-w844d\" (UID: \"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8\") " pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.371883 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrmbl\" (UniqueName: \"kubernetes.io/projected/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8-kube-api-access-hrmbl\") pod \"auto-csr-approver-29558002-w844d\" (UID: \"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8\") " pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.439148 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.652629 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558002-w844d"] Mar 14 09:22:00 crc kubenswrapper[4843]: I0314 09:22:00.942017 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558002-w844d" event={"ID":"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8","Type":"ContainerStarted","Data":"a3df4edfa29d4507515f83def75bf0ca23530b11e0b3f076748899baadaaccbc"} Mar 14 09:22:01 crc kubenswrapper[4843]: I0314 09:22:01.950679 4843 generic.go:334] "Generic (PLEG): container finished" podID="54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8" containerID="0686858794fd5e4868ec4687669da644f44286f475b98ae0fcbb07066f2f50de" exitCode=0 Mar 14 09:22:01 crc kubenswrapper[4843]: I0314 09:22:01.950927 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558002-w844d" event={"ID":"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8","Type":"ContainerDied","Data":"0686858794fd5e4868ec4687669da644f44286f475b98ae0fcbb07066f2f50de"} Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.216626 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.386639 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrmbl\" (UniqueName: \"kubernetes.io/projected/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8-kube-api-access-hrmbl\") pod \"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8\" (UID: \"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8\") " Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.392428 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8-kube-api-access-hrmbl" (OuterVolumeSpecName: "kube-api-access-hrmbl") pod "54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8" (UID: "54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8"). InnerVolumeSpecName "kube-api-access-hrmbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.488627 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrmbl\" (UniqueName: \"kubernetes.io/projected/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8-kube-api-access-hrmbl\") on node \"crc\" DevicePath \"\"" Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.962783 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558002-w844d" event={"ID":"54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8","Type":"ContainerDied","Data":"a3df4edfa29d4507515f83def75bf0ca23530b11e0b3f076748899baadaaccbc"} Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.962831 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3df4edfa29d4507515f83def75bf0ca23530b11e0b3f076748899baadaaccbc" Mar 14 09:22:03 crc kubenswrapper[4843]: I0314 09:22:03.962835 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558002-w844d" Mar 14 09:22:04 crc kubenswrapper[4843]: I0314 09:22:04.355246 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557996-tplp2"] Mar 14 09:22:04 crc kubenswrapper[4843]: I0314 09:22:04.358729 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557996-tplp2"] Mar 14 09:22:05 crc kubenswrapper[4843]: I0314 09:22:05.345897 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1108730f-bc5a-472b-98d5-087d6e04ad2a" path="/var/lib/kubelet/pods/1108730f-bc5a-472b-98d5-087d6e04ad2a/volumes" Mar 14 09:22:08 crc kubenswrapper[4843]: I0314 09:22:08.338060 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:22:08 crc kubenswrapper[4843]: I0314 09:22:08.338893 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" Mar 14 09:22:08 crc kubenswrapper[4843]: I0314 09:22:08.572480 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4"] Mar 14 09:22:08 crc kubenswrapper[4843]: I0314 09:22:08.986996 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" event={"ID":"35d96605-ef77-45e3-b12c-26c192a2dd50","Type":"ContainerStarted","Data":"52696615d62c6f1d13f2b42db7ecec57ea0979e0e1846bd14b0c2302630b776d"} Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.337921 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.337952 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.337962 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.338206 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.338547 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.338568 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.338918 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.338960 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.640636 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-8glvj"] Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.767321 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-7l9bs"] Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.772878 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5"] Mar 14 09:22:09 crc kubenswrapper[4843]: W0314 09:22:09.773318 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f706824_088d_4d70_830e_4844447b900d.slice/crio-6e50d136098b9a290278d470ae687facb3254ba8ecca03f5cd71b75549ccad98 WatchSource:0}: Error finding container 6e50d136098b9a290278d470ae687facb3254ba8ecca03f5cd71b75549ccad98: Status 404 returned error can't find the container with id 6e50d136098b9a290278d470ae687facb3254ba8ecca03f5cd71b75549ccad98 Mar 14 09:22:09 crc kubenswrapper[4843]: W0314 09:22:09.783564 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8bb5f31_a201_4fa9_a297_f294c2aefe07.slice/crio-3c0657a472d6b0f2b6648ae8bdf79c55ba90e4f7a160c61e44f835581f66b0cf WatchSource:0}: Error finding container 3c0657a472d6b0f2b6648ae8bdf79c55ba90e4f7a160c61e44f835581f66b0cf: Status 404 returned error can't find the container with id 3c0657a472d6b0f2b6648ae8bdf79c55ba90e4f7a160c61e44f835581f66b0cf Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.797346 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk"] Mar 14 09:22:09 crc kubenswrapper[4843]: W0314 09:22:09.804003 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d01a80d_2445_472e_ab28_81703fb66a27.slice/crio-b9e5939a74061a15a0925daaadad98af28574665f0fe85fb5daae61168b9617f WatchSource:0}: Error finding container b9e5939a74061a15a0925daaadad98af28574665f0fe85fb5daae61168b9617f: Status 404 returned error can't find the container with id b9e5939a74061a15a0925daaadad98af28574665f0fe85fb5daae61168b9617f Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.995523 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" event={"ID":"7d01a80d-2445-472e-ab28-81703fb66a27","Type":"ContainerStarted","Data":"b9e5939a74061a15a0925daaadad98af28574665f0fe85fb5daae61168b9617f"} Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.997263 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" event={"ID":"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c","Type":"ContainerStarted","Data":"d30ef3db84a319f07fd9467c549b23fb938d3a8045e4ce0f19719b41d756e219"} Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.998059 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" event={"ID":"7f706824-088d-4d70-830e-4844447b900d","Type":"ContainerStarted","Data":"6e50d136098b9a290278d470ae687facb3254ba8ecca03f5cd71b75549ccad98"} Mar 14 09:22:09 crc kubenswrapper[4843]: I0314 09:22:09.999141 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" event={"ID":"d8bb5f31-a201-4fa9-a297-f294c2aefe07","Type":"ContainerStarted","Data":"3c0657a472d6b0f2b6648ae8bdf79c55ba90e4f7a160c61e44f835581f66b0cf"} Mar 14 09:22:16 crc kubenswrapper[4843]: I0314 09:22:16.499337 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rqzbl" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.075361 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" event={"ID":"7f706824-088d-4d70-830e-4844447b900d","Type":"ContainerStarted","Data":"a7280f307aedd7f008ee9e3bd688936eaa9abdb4a808e7e2b684e3d37445f53b"} Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.077651 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" event={"ID":"d8bb5f31-a201-4fa9-a297-f294c2aefe07","Type":"ContainerStarted","Data":"283b4a2a297a00283025cdd7f984b085aa12661c81471f761ed568ceedf85d4a"} Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.077858 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.079626 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" event={"ID":"7d01a80d-2445-472e-ab28-81703fb66a27","Type":"ContainerStarted","Data":"cee11d107237b079f7489791fe202eaf33ec5336f172221a61ef709520468960"} Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.080129 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.081518 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" event={"ID":"35d96605-ef77-45e3-b12c-26c192a2dd50","Type":"ContainerStarted","Data":"231c05715cd9f5027a150bf82092b29431de28ffa6114e8b29b0ffe63309c42d"} Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.083037 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" event={"ID":"5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c","Type":"ContainerStarted","Data":"bce20401557507cff8a35a2b880025f5c1754eb7de077581f509c4ccd28e9840"} Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.083549 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.099005 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-7vsx5" podStartSLOduration=20.993584242 podStartE2EDuration="29.098989603s" podCreationTimestamp="2026-03-14 09:21:50 +0000 UTC" firstStartedPulling="2026-03-14 09:22:09.77691373 +0000 UTC m=+697.089524858" lastFinishedPulling="2026-03-14 09:22:17.882319091 +0000 UTC m=+705.194930219" observedRunningTime="2026-03-14 09:22:19.096814459 +0000 UTC m=+706.409425607" watchObservedRunningTime="2026-03-14 09:22:19.098989603 +0000 UTC m=+706.411600731" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.132677 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-7l9bs" podStartSLOduration=20.998719323 podStartE2EDuration="29.132655757s" podCreationTimestamp="2026-03-14 09:21:50 +0000 UTC" firstStartedPulling="2026-03-14 09:22:09.785930239 +0000 UTC m=+697.098541367" lastFinishedPulling="2026-03-14 09:22:17.919866683 +0000 UTC m=+705.232477801" observedRunningTime="2026-03-14 09:22:19.12844899 +0000 UTC m=+706.441060128" watchObservedRunningTime="2026-03-14 09:22:19.132655757 +0000 UTC m=+706.445266885" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.162431 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-scrd4" podStartSLOduration=19.837430194 podStartE2EDuration="29.162412832s" podCreationTimestamp="2026-03-14 09:21:50 +0000 UTC" firstStartedPulling="2026-03-14 09:22:08.579580277 +0000 UTC m=+695.892191405" lastFinishedPulling="2026-03-14 09:22:17.904562915 +0000 UTC m=+705.217174043" observedRunningTime="2026-03-14 09:22:19.160480242 +0000 UTC m=+706.473091610" watchObservedRunningTime="2026-03-14 09:22:19.162412832 +0000 UTC m=+706.475023960" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.183027 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" podStartSLOduration=21.006501869 podStartE2EDuration="29.183012753s" podCreationTimestamp="2026-03-14 09:21:50 +0000 UTC" firstStartedPulling="2026-03-14 09:22:09.705706574 +0000 UTC m=+697.018317712" lastFinishedPulling="2026-03-14 09:22:17.882217458 +0000 UTC m=+705.194828596" observedRunningTime="2026-03-14 09:22:19.181437584 +0000 UTC m=+706.494048722" watchObservedRunningTime="2026-03-14 09:22:19.183012753 +0000 UTC m=+706.495623881" Mar 14 09:22:19 crc kubenswrapper[4843]: I0314 09:22:19.203358 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-69fb744b97-5knkk" podStartSLOduration=21.130199816 podStartE2EDuration="29.203338089s" podCreationTimestamp="2026-03-14 09:21:50 +0000 UTC" firstStartedPulling="2026-03-14 09:22:09.808825989 +0000 UTC m=+697.121437117" lastFinishedPulling="2026-03-14 09:22:17.881964262 +0000 UTC m=+705.194575390" observedRunningTime="2026-03-14 09:22:19.202654561 +0000 UTC m=+706.515265709" watchObservedRunningTime="2026-03-14 09:22:19.203338089 +0000 UTC m=+706.515949217" Mar 14 09:22:27 crc kubenswrapper[4843]: I0314 09:22:27.964764 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2"] Mar 14 09:22:27 crc kubenswrapper[4843]: E0314 09:22:27.965886 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8" containerName="oc" Mar 14 09:22:27 crc kubenswrapper[4843]: I0314 09:22:27.965902 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8" containerName="oc" Mar 14 09:22:27 crc kubenswrapper[4843]: I0314 09:22:27.966044 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8" containerName="oc" Mar 14 09:22:27 crc kubenswrapper[4843]: I0314 09:22:27.967119 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:27 crc kubenswrapper[4843]: I0314 09:22:27.969245 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 14 09:22:27 crc kubenswrapper[4843]: I0314 09:22:27.976448 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2"] Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.112610 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.112687 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22knm\" (UniqueName: \"kubernetes.io/projected/e952e216-1adb-456a-9f7e-61d2dc2ddb78-kube-api-access-22knm\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.112755 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.214737 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.215031 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.215036 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.215137 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22knm\" (UniqueName: \"kubernetes.io/projected/e952e216-1adb-456a-9f7e-61d2dc2ddb78-kube-api-access-22knm\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.215507 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.240106 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22knm\" (UniqueName: \"kubernetes.io/projected/e952e216-1adb-456a-9f7e-61d2dc2ddb78-kube-api-access-22knm\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.284912 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:28 crc kubenswrapper[4843]: I0314 09:22:28.489112 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2"] Mar 14 09:22:28 crc kubenswrapper[4843]: W0314 09:22:28.503036 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode952e216_1adb_456a_9f7e_61d2dc2ddb78.slice/crio-2ca267b25b59f0a30f97c8062f35912f1755a30db02d4a7284101a15ec451168 WatchSource:0}: Error finding container 2ca267b25b59f0a30f97c8062f35912f1755a30db02d4a7284101a15ec451168: Status 404 returned error can't find the container with id 2ca267b25b59f0a30f97c8062f35912f1755a30db02d4a7284101a15ec451168 Mar 14 09:22:29 crc kubenswrapper[4843]: I0314 09:22:29.140210 4843 generic.go:334] "Generic (PLEG): container finished" podID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerID="b80a71428fa37ddba8cf68be6dcfce083544dd27b1a6f1c3f6329231318c86de" exitCode=0 Mar 14 09:22:29 crc kubenswrapper[4843]: I0314 09:22:29.140449 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" event={"ID":"e952e216-1adb-456a-9f7e-61d2dc2ddb78","Type":"ContainerDied","Data":"b80a71428fa37ddba8cf68be6dcfce083544dd27b1a6f1c3f6329231318c86de"} Mar 14 09:22:29 crc kubenswrapper[4843]: I0314 09:22:29.140474 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" event={"ID":"e952e216-1adb-456a-9f7e-61d2dc2ddb78","Type":"ContainerStarted","Data":"2ca267b25b59f0a30f97c8062f35912f1755a30db02d4a7284101a15ec451168"} Mar 14 09:22:30 crc kubenswrapper[4843]: I0314 09:22:30.830960 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-8glvj" Mar 14 09:22:31 crc kubenswrapper[4843]: I0314 09:22:31.152749 4843 generic.go:334] "Generic (PLEG): container finished" podID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerID="4178e1770d65b4eb211db04277082ecccd5552e10965cd6373a8a593d8e97693" exitCode=0 Mar 14 09:22:31 crc kubenswrapper[4843]: I0314 09:22:31.152791 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" event={"ID":"e952e216-1adb-456a-9f7e-61d2dc2ddb78","Type":"ContainerDied","Data":"4178e1770d65b4eb211db04277082ecccd5552e10965cd6373a8a593d8e97693"} Mar 14 09:22:32 crc kubenswrapper[4843]: I0314 09:22:32.164771 4843 generic.go:334] "Generic (PLEG): container finished" podID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerID="9dd46c1df00ab19415642ce469c7143021e272e949f35b915ddf8d18a0502a32" exitCode=0 Mar 14 09:22:32 crc kubenswrapper[4843]: I0314 09:22:32.164808 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" event={"ID":"e952e216-1adb-456a-9f7e-61d2dc2ddb78","Type":"ContainerDied","Data":"9dd46c1df00ab19415642ce469c7143021e272e949f35b915ddf8d18a0502a32"} Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.417264 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.596140 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-util\") pod \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.596198 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22knm\" (UniqueName: \"kubernetes.io/projected/e952e216-1adb-456a-9f7e-61d2dc2ddb78-kube-api-access-22knm\") pod \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.596255 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-bundle\") pod \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\" (UID: \"e952e216-1adb-456a-9f7e-61d2dc2ddb78\") " Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.596802 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-bundle" (OuterVolumeSpecName: "bundle") pod "e952e216-1adb-456a-9f7e-61d2dc2ddb78" (UID: "e952e216-1adb-456a-9f7e-61d2dc2ddb78"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.612457 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-util" (OuterVolumeSpecName: "util") pod "e952e216-1adb-456a-9f7e-61d2dc2ddb78" (UID: "e952e216-1adb-456a-9f7e-61d2dc2ddb78"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.612949 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e952e216-1adb-456a-9f7e-61d2dc2ddb78-kube-api-access-22knm" (OuterVolumeSpecName: "kube-api-access-22knm") pod "e952e216-1adb-456a-9f7e-61d2dc2ddb78" (UID: "e952e216-1adb-456a-9f7e-61d2dc2ddb78"). InnerVolumeSpecName "kube-api-access-22knm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.697290 4843 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-util\") on node \"crc\" DevicePath \"\"" Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.697322 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22knm\" (UniqueName: \"kubernetes.io/projected/e952e216-1adb-456a-9f7e-61d2dc2ddb78-kube-api-access-22knm\") on node \"crc\" DevicePath \"\"" Mar 14 09:22:33 crc kubenswrapper[4843]: I0314 09:22:33.697337 4843 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e952e216-1adb-456a-9f7e-61d2dc2ddb78-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:22:34 crc kubenswrapper[4843]: I0314 09:22:34.177997 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" event={"ID":"e952e216-1adb-456a-9f7e-61d2dc2ddb78","Type":"ContainerDied","Data":"2ca267b25b59f0a30f97c8062f35912f1755a30db02d4a7284101a15ec451168"} Mar 14 09:22:34 crc kubenswrapper[4843]: I0314 09:22:34.178038 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2" Mar 14 09:22:34 crc kubenswrapper[4843]: I0314 09:22:34.178037 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ca267b25b59f0a30f97c8062f35912f1755a30db02d4a7284101a15ec451168" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.655900 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d"] Mar 14 09:22:39 crc kubenswrapper[4843]: E0314 09:22:39.656712 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="extract" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.656729 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="extract" Mar 14 09:22:39 crc kubenswrapper[4843]: E0314 09:22:39.656744 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="util" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.656753 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="util" Mar 14 09:22:39 crc kubenswrapper[4843]: E0314 09:22:39.656767 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="pull" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.656775 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="pull" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.656900 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="e952e216-1adb-456a-9f7e-61d2dc2ddb78" containerName="extract" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.657417 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.659518 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-9dhtw" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.659539 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.659562 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.665341 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v879s\" (UniqueName: \"kubernetes.io/projected/63096d1d-7063-4821-b735-e3e65e2b9f64-kube-api-access-v879s\") pod \"nmstate-operator-796d4cfff4-n7k5d\" (UID: \"63096d1d-7063-4821-b735-e3e65e2b9f64\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.707884 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d"] Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.766778 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v879s\" (UniqueName: \"kubernetes.io/projected/63096d1d-7063-4821-b735-e3e65e2b9f64-kube-api-access-v879s\") pod \"nmstate-operator-796d4cfff4-n7k5d\" (UID: \"63096d1d-7063-4821-b735-e3e65e2b9f64\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.787366 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v879s\" (UniqueName: \"kubernetes.io/projected/63096d1d-7063-4821-b735-e3e65e2b9f64-kube-api-access-v879s\") pod \"nmstate-operator-796d4cfff4-n7k5d\" (UID: \"63096d1d-7063-4821-b735-e3e65e2b9f64\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" Mar 14 09:22:39 crc kubenswrapper[4843]: I0314 09:22:39.974099 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" Mar 14 09:22:40 crc kubenswrapper[4843]: I0314 09:22:40.211305 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d"] Mar 14 09:22:40 crc kubenswrapper[4843]: W0314 09:22:40.223571 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63096d1d_7063_4821_b735_e3e65e2b9f64.slice/crio-a1d11fb41f9f0d63d6cd4b3f7adec53a2dd92a3da3a515e22589eff57e4f7f00 WatchSource:0}: Error finding container a1d11fb41f9f0d63d6cd4b3f7adec53a2dd92a3da3a515e22589eff57e4f7f00: Status 404 returned error can't find the container with id a1d11fb41f9f0d63d6cd4b3f7adec53a2dd92a3da3a515e22589eff57e4f7f00 Mar 14 09:22:40 crc kubenswrapper[4843]: I0314 09:22:40.344217 4843 scope.go:117] "RemoveContainer" containerID="641bd9b39d70fdb98f7edbd95c7636b645a2fa7fdb0224d8776d9c156f24f6f8" Mar 14 09:22:41 crc kubenswrapper[4843]: I0314 09:22:41.220313 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" event={"ID":"63096d1d-7063-4821-b735-e3e65e2b9f64","Type":"ContainerStarted","Data":"a1d11fb41f9f0d63d6cd4b3f7adec53a2dd92a3da3a515e22589eff57e4f7f00"} Mar 14 09:22:42 crc kubenswrapper[4843]: I0314 09:22:42.073338 4843 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 14 09:22:44 crc kubenswrapper[4843]: I0314 09:22:44.239990 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" event={"ID":"63096d1d-7063-4821-b735-e3e65e2b9f64","Type":"ContainerStarted","Data":"1bd75d30b3f53293a5448d1127f0c28823c9c9699b75ea2beb5ace502449c504"} Mar 14 09:22:44 crc kubenswrapper[4843]: I0314 09:22:44.261096 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-n7k5d" podStartSLOduration=2.125587409 podStartE2EDuration="5.261071241s" podCreationTimestamp="2026-03-14 09:22:39 +0000 UTC" firstStartedPulling="2026-03-14 09:22:40.225178578 +0000 UTC m=+727.537789706" lastFinishedPulling="2026-03-14 09:22:43.36066242 +0000 UTC m=+730.673273538" observedRunningTime="2026-03-14 09:22:44.256130697 +0000 UTC m=+731.568741835" watchObservedRunningTime="2026-03-14 09:22:44.261071241 +0000 UTC m=+731.573682369" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.178641 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.181692 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.187142 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-p42gc" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.222329 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.228029 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.228941 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.233373 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fqs5c"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.234476 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.239164 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.242178 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.289545 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtk6p\" (UniqueName: \"kubernetes.io/projected/c108acb5-004b-4c4e-b8e9-44aed45a2b86-kube-api-access-mtk6p\") pod \"nmstate-metrics-9b8c8685d-bcwrg\" (UID: \"c108acb5-004b-4c4e-b8e9-44aed45a2b86\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.334588 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.335371 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.337604 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-b9p5n" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.345199 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.345205 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.358488 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391041 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kskct\" (UniqueName: \"kubernetes.io/projected/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-kube-api-access-kskct\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391093 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/185eedfb-940a-4fb8-943f-3f2761f44301-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hcdc6\" (UID: \"185eedfb-940a-4fb8-943f-3f2761f44301\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391497 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-dbus-socket\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391528 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-nmstate-lock\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391544 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfv9n\" (UniqueName: \"kubernetes.io/projected/185eedfb-940a-4fb8-943f-3f2761f44301-kube-api-access-sfv9n\") pod \"nmstate-webhook-5f558f5558-hcdc6\" (UID: \"185eedfb-940a-4fb8-943f-3f2761f44301\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391652 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtk6p\" (UniqueName: \"kubernetes.io/projected/c108acb5-004b-4c4e-b8e9-44aed45a2b86-kube-api-access-mtk6p\") pod \"nmstate-metrics-9b8c8685d-bcwrg\" (UID: \"c108acb5-004b-4c4e-b8e9-44aed45a2b86\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.391693 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-ovs-socket\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.415155 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtk6p\" (UniqueName: \"kubernetes.io/projected/c108acb5-004b-4c4e-b8e9-44aed45a2b86-kube-api-access-mtk6p\") pod \"nmstate-metrics-9b8c8685d-bcwrg\" (UID: \"c108acb5-004b-4c4e-b8e9-44aed45a2b86\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.492553 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/22c4bb4f-2852-4c58-a0db-17793ec80671-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.492656 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-ovs-socket\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.492702 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/22c4bb4f-2852-4c58-a0db-17793ec80671-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.492772 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kskct\" (UniqueName: \"kubernetes.io/projected/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-kube-api-access-kskct\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.492790 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-ovs-socket\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.492799 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/185eedfb-940a-4fb8-943f-3f2761f44301-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hcdc6\" (UID: \"185eedfb-940a-4fb8-943f-3f2761f44301\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.493452 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlgl6\" (UniqueName: \"kubernetes.io/projected/22c4bb4f-2852-4c58-a0db-17793ec80671-kube-api-access-tlgl6\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.493532 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-dbus-socket\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.493564 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-nmstate-lock\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.493653 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfv9n\" (UniqueName: \"kubernetes.io/projected/185eedfb-940a-4fb8-943f-3f2761f44301-kube-api-access-sfv9n\") pod \"nmstate-webhook-5f558f5558-hcdc6\" (UID: \"185eedfb-940a-4fb8-943f-3f2761f44301\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.493606 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-nmstate-lock\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.493740 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-dbus-socket\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.504880 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/185eedfb-940a-4fb8-943f-3f2761f44301-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hcdc6\" (UID: \"185eedfb-940a-4fb8-943f-3f2761f44301\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.506724 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.526409 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfv9n\" (UniqueName: \"kubernetes.io/projected/185eedfb-940a-4fb8-943f-3f2761f44301-kube-api-access-sfv9n\") pod \"nmstate-webhook-5f558f5558-hcdc6\" (UID: \"185eedfb-940a-4fb8-943f-3f2761f44301\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.528239 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kskct\" (UniqueName: \"kubernetes.io/projected/951a69ab-6e2a-4ff3-bcf3-db69fbfcb349-kube-api-access-kskct\") pod \"nmstate-handler-fqs5c\" (UID: \"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349\") " pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.532745 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-584b867db4-62h25"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.533448 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.543258 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-584b867db4-62h25"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.552416 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.562209 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595100 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/22c4bb4f-2852-4c58-a0db-17793ec80671-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595146 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-service-ca\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595191 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlgl6\" (UniqueName: \"kubernetes.io/projected/22c4bb4f-2852-4c58-a0db-17793ec80671-kube-api-access-tlgl6\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595209 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-oauth-serving-cert\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595339 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-serving-cert\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595404 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-oauth-config\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595457 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-trusted-ca-bundle\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595516 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/22c4bb4f-2852-4c58-a0db-17793ec80671-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595563 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-config\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.595587 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nghsq\" (UniqueName: \"kubernetes.io/projected/ab35bbcf-2757-4b2e-8a4d-7a716669c229-kube-api-access-nghsq\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.596548 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/22c4bb4f-2852-4c58-a0db-17793ec80671-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.598665 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/22c4bb4f-2852-4c58-a0db-17793ec80671-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.615828 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlgl6\" (UniqueName: \"kubernetes.io/projected/22c4bb4f-2852-4c58-a0db-17793ec80671-kube-api-access-tlgl6\") pod \"nmstate-console-plugin-86f58fcf4-wwth2\" (UID: \"22c4bb4f-2852-4c58-a0db-17793ec80671\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.657519 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699106 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-config\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699411 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nghsq\" (UniqueName: \"kubernetes.io/projected/ab35bbcf-2757-4b2e-8a4d-7a716669c229-kube-api-access-nghsq\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699440 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-service-ca\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699460 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-oauth-serving-cert\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699480 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-serving-cert\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699495 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-oauth-config\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.699522 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-trusted-ca-bundle\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.700488 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-config\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.701094 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-oauth-serving-cert\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.701621 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-service-ca\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.701949 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-trusted-ca-bundle\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.707258 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-oauth-config\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.707341 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-serving-cert\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.726088 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nghsq\" (UniqueName: \"kubernetes.io/projected/ab35bbcf-2757-4b2e-8a4d-7a716669c229-kube-api-access-nghsq\") pod \"console-584b867db4-62h25\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.758186 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.815256 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6"] Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.893426 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2"] Mar 14 09:22:49 crc kubenswrapper[4843]: W0314 09:22:49.896309 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22c4bb4f_2852_4c58_a0db_17793ec80671.slice/crio-c3ef2169d187af2aa0c34ac98b00145df549fded355a2cc689144609cdf2dfbf WatchSource:0}: Error finding container c3ef2169d187af2aa0c34ac98b00145df549fded355a2cc689144609cdf2dfbf: Status 404 returned error can't find the container with id c3ef2169d187af2aa0c34ac98b00145df549fded355a2cc689144609cdf2dfbf Mar 14 09:22:49 crc kubenswrapper[4843]: I0314 09:22:49.918714 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:50 crc kubenswrapper[4843]: I0314 09:22:50.089083 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-584b867db4-62h25"] Mar 14 09:22:50 crc kubenswrapper[4843]: W0314 09:22:50.094376 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab35bbcf_2757_4b2e_8a4d_7a716669c229.slice/crio-f8b38cb17dd8d4896cb805c622dca520de6f37f7fd8a1aca944cb15ead624fa9 WatchSource:0}: Error finding container f8b38cb17dd8d4896cb805c622dca520de6f37f7fd8a1aca944cb15ead624fa9: Status 404 returned error can't find the container with id f8b38cb17dd8d4896cb805c622dca520de6f37f7fd8a1aca944cb15ead624fa9 Mar 14 09:22:50 crc kubenswrapper[4843]: I0314 09:22:50.274588 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-584b867db4-62h25" event={"ID":"ab35bbcf-2757-4b2e-8a4d-7a716669c229","Type":"ContainerStarted","Data":"f8b38cb17dd8d4896cb805c622dca520de6f37f7fd8a1aca944cb15ead624fa9"} Mar 14 09:22:50 crc kubenswrapper[4843]: I0314 09:22:50.275727 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" event={"ID":"22c4bb4f-2852-4c58-a0db-17793ec80671","Type":"ContainerStarted","Data":"c3ef2169d187af2aa0c34ac98b00145df549fded355a2cc689144609cdf2dfbf"} Mar 14 09:22:50 crc kubenswrapper[4843]: I0314 09:22:50.277083 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" event={"ID":"185eedfb-940a-4fb8-943f-3f2761f44301","Type":"ContainerStarted","Data":"c28ad5dc05c2ed3698ad027596adb98fd581ef6b624bad51a68272f0a32dc6aa"} Mar 14 09:22:50 crc kubenswrapper[4843]: I0314 09:22:50.277981 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fqs5c" event={"ID":"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349","Type":"ContainerStarted","Data":"1bc7fcd8dcd07836043f6b4ec2087cc283fc81da070f511792c707e2d5ec8fe6"} Mar 14 09:22:50 crc kubenswrapper[4843]: I0314 09:22:50.278885 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" event={"ID":"c108acb5-004b-4c4e-b8e9-44aed45a2b86","Type":"ContainerStarted","Data":"f7d8a56ee3100ca0d0847c21ab7b826eec02bf37534647755664adb97185d2db"} Mar 14 09:22:52 crc kubenswrapper[4843]: I0314 09:22:52.295351 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-584b867db4-62h25" event={"ID":"ab35bbcf-2757-4b2e-8a4d-7a716669c229","Type":"ContainerStarted","Data":"7759b174c9c43bdc8815200b690a32544cb33e2859bb7f6e40d327e1baad80a4"} Mar 14 09:22:53 crc kubenswrapper[4843]: I0314 09:22:53.326012 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-584b867db4-62h25" podStartSLOduration=4.325987836 podStartE2EDuration="4.325987836s" podCreationTimestamp="2026-03-14 09:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:22:53.318521236 +0000 UTC m=+740.631132384" watchObservedRunningTime="2026-03-14 09:22:53.325987836 +0000 UTC m=+740.638598964" Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.325496 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fqs5c" event={"ID":"951a69ab-6e2a-4ff3-bcf3-db69fbfcb349","Type":"ContainerStarted","Data":"b45ac241de0274ae98584a34ae07ac624fd96ce95eda0fb7beec5918df4156a3"} Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.326145 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.326991 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" event={"ID":"c108acb5-004b-4c4e-b8e9-44aed45a2b86","Type":"ContainerStarted","Data":"773918f21e5e4b504492eade639341a1d5e9e2ebeba21e5304fbee5f869107a9"} Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.328902 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" event={"ID":"22c4bb4f-2852-4c58-a0db-17793ec80671","Type":"ContainerStarted","Data":"927e557564e58ab600da05c61ffb882cbc6d46ae6cb8b8b3640142f3d53977f6"} Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.332901 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" event={"ID":"185eedfb-940a-4fb8-943f-3f2761f44301","Type":"ContainerStarted","Data":"6ea9cbf8525e21f1e714d94eed26e598a81cb8010083d0977439261487e0c305"} Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.333780 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.379288 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fqs5c" podStartSLOduration=1.262733399 podStartE2EDuration="7.379228649s" podCreationTimestamp="2026-03-14 09:22:49 +0000 UTC" firstStartedPulling="2026-03-14 09:22:49.601957555 +0000 UTC m=+736.914568683" lastFinishedPulling="2026-03-14 09:22:55.718452805 +0000 UTC m=+743.031063933" observedRunningTime="2026-03-14 09:22:56.352658897 +0000 UTC m=+743.665270055" watchObservedRunningTime="2026-03-14 09:22:56.379228649 +0000 UTC m=+743.691839797" Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.384475 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wwth2" podStartSLOduration=2.16431653 podStartE2EDuration="7.384447881s" podCreationTimestamp="2026-03-14 09:22:49 +0000 UTC" firstStartedPulling="2026-03-14 09:22:49.898525292 +0000 UTC m=+737.211136430" lastFinishedPulling="2026-03-14 09:22:55.118656613 +0000 UTC m=+742.431267781" observedRunningTime="2026-03-14 09:22:56.370409836 +0000 UTC m=+743.683021004" watchObservedRunningTime="2026-03-14 09:22:56.384447881 +0000 UTC m=+743.697059019" Mar 14 09:22:56 crc kubenswrapper[4843]: I0314 09:22:56.407029 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" podStartSLOduration=1.454027944 podStartE2EDuration="7.407004751s" podCreationTimestamp="2026-03-14 09:22:49 +0000 UTC" firstStartedPulling="2026-03-14 09:22:49.825981858 +0000 UTC m=+737.138592986" lastFinishedPulling="2026-03-14 09:22:55.778958665 +0000 UTC m=+743.091569793" observedRunningTime="2026-03-14 09:22:56.405956274 +0000 UTC m=+743.718567402" watchObservedRunningTime="2026-03-14 09:22:56.407004751 +0000 UTC m=+743.719615899" Mar 14 09:22:59 crc kubenswrapper[4843]: I0314 09:22:59.918851 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:59 crc kubenswrapper[4843]: I0314 09:22:59.919567 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:22:59 crc kubenswrapper[4843]: I0314 09:22:59.924944 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:23:00 crc kubenswrapper[4843]: I0314 09:23:00.286455 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" event={"ID":"c108acb5-004b-4c4e-b8e9-44aed45a2b86","Type":"ContainerStarted","Data":"f07887e9eb1fecc676c64bcd2f7cae0e02a96aec881f6929de353cfb87aaf06d"} Mar 14 09:23:00 crc kubenswrapper[4843]: I0314 09:23:00.290325 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:23:00 crc kubenswrapper[4843]: I0314 09:23:00.304090 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-bcwrg" podStartSLOduration=1.339737335 podStartE2EDuration="11.304068875s" podCreationTimestamp="2026-03-14 09:22:49 +0000 UTC" firstStartedPulling="2026-03-14 09:22:49.771793488 +0000 UTC m=+737.084404616" lastFinishedPulling="2026-03-14 09:22:59.736125028 +0000 UTC m=+747.048736156" observedRunningTime="2026-03-14 09:23:00.301263534 +0000 UTC m=+747.613874662" watchObservedRunningTime="2026-03-14 09:23:00.304068875 +0000 UTC m=+747.616680003" Mar 14 09:23:00 crc kubenswrapper[4843]: I0314 09:23:00.411931 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2j8wl"] Mar 14 09:23:04 crc kubenswrapper[4843]: I0314 09:23:04.590854 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fqs5c" Mar 14 09:23:09 crc kubenswrapper[4843]: I0314 09:23:09.561582 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hcdc6" Mar 14 09:23:15 crc kubenswrapper[4843]: I0314 09:23:15.387415 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:23:15 crc kubenswrapper[4843]: I0314 09:23:15.387958 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.065322 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429"] Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.067194 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.073060 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.079154 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429"] Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.200435 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.200539 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.200617 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj69g\" (UniqueName: \"kubernetes.io/projected/d916f254-7cc8-48b9-b01e-ba6270584eb3-kube-api-access-cj69g\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.301849 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj69g\" (UniqueName: \"kubernetes.io/projected/d916f254-7cc8-48b9-b01e-ba6270584eb3-kube-api-access-cj69g\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.301983 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.302020 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.302566 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.302636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.322428 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj69g\" (UniqueName: \"kubernetes.io/projected/d916f254-7cc8-48b9-b01e-ba6270584eb3-kube-api-access-cj69g\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.390506 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:24 crc kubenswrapper[4843]: I0314 09:23:24.599578 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429"] Mar 14 09:23:25 crc kubenswrapper[4843]: I0314 09:23:25.451930 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-2j8wl" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerName="console" containerID="cri-o://9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4" gracePeriod=15 Mar 14 09:23:25 crc kubenswrapper[4843]: I0314 09:23:25.483175 4843 generic.go:334] "Generic (PLEG): container finished" podID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerID="5db35385a3f993d6a905e90b3e28e64571caaf597a8194a8c4afa23c787db61b" exitCode=0 Mar 14 09:23:25 crc kubenswrapper[4843]: I0314 09:23:25.483455 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" event={"ID":"d916f254-7cc8-48b9-b01e-ba6270584eb3","Type":"ContainerDied","Data":"5db35385a3f993d6a905e90b3e28e64571caaf597a8194a8c4afa23c787db61b"} Mar 14 09:23:25 crc kubenswrapper[4843]: I0314 09:23:25.483603 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" event={"ID":"d916f254-7cc8-48b9-b01e-ba6270584eb3","Type":"ContainerStarted","Data":"db4ea3a5b68584d75a30107e60b67574f4cb841e4ba4ae9b290df5ffc2a9c904"} Mar 14 09:23:25 crc kubenswrapper[4843]: I0314 09:23:25.485747 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.349218 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2j8wl_5c05075c-f721-4e78-ad5e-aaf88989336d/console/0.log" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.349314 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.413505 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zhsnn"] Mar 14 09:23:26 crc kubenswrapper[4843]: E0314 09:23:26.413783 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerName="console" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.413802 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerName="console" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.413949 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerName="console" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.415468 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.423603 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zhsnn"] Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428495 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-service-ca\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428555 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-serving-cert\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428585 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-oauth-config\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428615 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ljjp\" (UniqueName: \"kubernetes.io/projected/5c05075c-f721-4e78-ad5e-aaf88989336d-kube-api-access-5ljjp\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428643 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-console-config\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428698 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-oauth-serving-cert\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428723 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-trusted-ca-bundle\") pod \"5c05075c-f721-4e78-ad5e-aaf88989336d\" (UID: \"5c05075c-f721-4e78-ad5e-aaf88989336d\") " Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428817 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-catalog-content\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.428874 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz7rn\" (UniqueName: \"kubernetes.io/projected/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-kube-api-access-zz7rn\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.429035 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-utilities\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.429450 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-console-config" (OuterVolumeSpecName: "console-config") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.429475 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.429724 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-service-ca" (OuterVolumeSpecName: "service-ca") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.429915 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.438075 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.443918 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.444047 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c05075c-f721-4e78-ad5e-aaf88989336d-kube-api-access-5ljjp" (OuterVolumeSpecName: "kube-api-access-5ljjp") pod "5c05075c-f721-4e78-ad5e-aaf88989336d" (UID: "5c05075c-f721-4e78-ad5e-aaf88989336d"). InnerVolumeSpecName "kube-api-access-5ljjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.491306 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2j8wl_5c05075c-f721-4e78-ad5e-aaf88989336d/console/0.log" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.491606 4843 generic.go:334] "Generic (PLEG): container finished" podID="5c05075c-f721-4e78-ad5e-aaf88989336d" containerID="9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4" exitCode=2 Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.491634 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2j8wl" event={"ID":"5c05075c-f721-4e78-ad5e-aaf88989336d","Type":"ContainerDied","Data":"9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4"} Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.491663 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2j8wl" event={"ID":"5c05075c-f721-4e78-ad5e-aaf88989336d","Type":"ContainerDied","Data":"9766f82fbbc308f68ee4c376c9e33a6cb44cc45b9a50df672f215295a9ad922d"} Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.491683 4843 scope.go:117] "RemoveContainer" containerID="9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.491789 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2j8wl" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.510079 4843 scope.go:117] "RemoveContainer" containerID="9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4" Mar 14 09:23:26 crc kubenswrapper[4843]: E0314 09:23:26.510862 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4\": container with ID starting with 9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4 not found: ID does not exist" containerID="9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.510911 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4"} err="failed to get container status \"9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4\": rpc error: code = NotFound desc = could not find container \"9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4\": container with ID starting with 9e6a89d1bbefa1fcb167514b9127405823726d3e279fd76e33c01627582e9de4 not found: ID does not exist" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.523405 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2j8wl"] Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.529284 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-2j8wl"] Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530048 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-utilities\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530118 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-catalog-content\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530173 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz7rn\" (UniqueName: \"kubernetes.io/projected/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-kube-api-access-zz7rn\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530236 4843 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530259 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530288 4843 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530301 4843 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530312 4843 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5c05075c-f721-4e78-ad5e-aaf88989336d-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530320 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ljjp\" (UniqueName: \"kubernetes.io/projected/5c05075c-f721-4e78-ad5e-aaf88989336d-kube-api-access-5ljjp\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530338 4843 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5c05075c-f721-4e78-ad5e-aaf88989336d-console-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530473 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-utilities\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.530497 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-catalog-content\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.551728 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz7rn\" (UniqueName: \"kubernetes.io/projected/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-kube-api-access-zz7rn\") pod \"redhat-operators-zhsnn\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.732230 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:26 crc kubenswrapper[4843]: I0314 09:23:26.975468 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zhsnn"] Mar 14 09:23:27 crc kubenswrapper[4843]: I0314 09:23:27.347794 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c05075c-f721-4e78-ad5e-aaf88989336d" path="/var/lib/kubelet/pods/5c05075c-f721-4e78-ad5e-aaf88989336d/volumes" Mar 14 09:23:27 crc kubenswrapper[4843]: I0314 09:23:27.500395 4843 generic.go:334] "Generic (PLEG): container finished" podID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerID="23964aa1bc27039b1d7e12398954c12fa6ad87342aadd11fcb88d42915fb5320" exitCode=0 Mar 14 09:23:27 crc kubenswrapper[4843]: I0314 09:23:27.500458 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" event={"ID":"d916f254-7cc8-48b9-b01e-ba6270584eb3","Type":"ContainerDied","Data":"23964aa1bc27039b1d7e12398954c12fa6ad87342aadd11fcb88d42915fb5320"} Mar 14 09:23:27 crc kubenswrapper[4843]: I0314 09:23:27.502467 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerID="ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de" exitCode=0 Mar 14 09:23:27 crc kubenswrapper[4843]: I0314 09:23:27.502505 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerDied","Data":"ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de"} Mar 14 09:23:27 crc kubenswrapper[4843]: I0314 09:23:27.502527 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerStarted","Data":"73967c6d445069267e3f6f9cf56a1eaa900a5c61b02c43228f1affe82b4d2dac"} Mar 14 09:23:28 crc kubenswrapper[4843]: I0314 09:23:28.509497 4843 generic.go:334] "Generic (PLEG): container finished" podID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerID="300c1b1345e5f3d54ab3348e04a04ccfe4a33277145bffca09164b56bf611c5c" exitCode=0 Mar 14 09:23:28 crc kubenswrapper[4843]: I0314 09:23:28.509576 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" event={"ID":"d916f254-7cc8-48b9-b01e-ba6270584eb3","Type":"ContainerDied","Data":"300c1b1345e5f3d54ab3348e04a04ccfe4a33277145bffca09164b56bf611c5c"} Mar 14 09:23:28 crc kubenswrapper[4843]: I0314 09:23:28.511607 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerStarted","Data":"7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789"} Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.522791 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerID="7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789" exitCode=0 Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.522838 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerDied","Data":"7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789"} Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.825752 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.976355 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-bundle\") pod \"d916f254-7cc8-48b9-b01e-ba6270584eb3\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.976473 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj69g\" (UniqueName: \"kubernetes.io/projected/d916f254-7cc8-48b9-b01e-ba6270584eb3-kube-api-access-cj69g\") pod \"d916f254-7cc8-48b9-b01e-ba6270584eb3\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.976529 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-util\") pod \"d916f254-7cc8-48b9-b01e-ba6270584eb3\" (UID: \"d916f254-7cc8-48b9-b01e-ba6270584eb3\") " Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.978072 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-bundle" (OuterVolumeSpecName: "bundle") pod "d916f254-7cc8-48b9-b01e-ba6270584eb3" (UID: "d916f254-7cc8-48b9-b01e-ba6270584eb3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:23:29 crc kubenswrapper[4843]: I0314 09:23:29.984236 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d916f254-7cc8-48b9-b01e-ba6270584eb3-kube-api-access-cj69g" (OuterVolumeSpecName: "kube-api-access-cj69g") pod "d916f254-7cc8-48b9-b01e-ba6270584eb3" (UID: "d916f254-7cc8-48b9-b01e-ba6270584eb3"). InnerVolumeSpecName "kube-api-access-cj69g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.006391 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-util" (OuterVolumeSpecName: "util") pod "d916f254-7cc8-48b9-b01e-ba6270584eb3" (UID: "d916f254-7cc8-48b9-b01e-ba6270584eb3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.077724 4843 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.077785 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj69g\" (UniqueName: \"kubernetes.io/projected/d916f254-7cc8-48b9-b01e-ba6270584eb3-kube-api-access-cj69g\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.077800 4843 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d916f254-7cc8-48b9-b01e-ba6270584eb3-util\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.531931 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" event={"ID":"d916f254-7cc8-48b9-b01e-ba6270584eb3","Type":"ContainerDied","Data":"db4ea3a5b68584d75a30107e60b67574f4cb841e4ba4ae9b290df5ffc2a9c904"} Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.531967 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.531985 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db4ea3a5b68584d75a30107e60b67574f4cb841e4ba4ae9b290df5ffc2a9c904" Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.534960 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerStarted","Data":"b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e"} Mar 14 09:23:30 crc kubenswrapper[4843]: I0314 09:23:30.565704 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zhsnn" podStartSLOduration=2.090954815 podStartE2EDuration="4.565681134s" podCreationTimestamp="2026-03-14 09:23:26 +0000 UTC" firstStartedPulling="2026-03-14 09:23:27.504219203 +0000 UTC m=+774.816830331" lastFinishedPulling="2026-03-14 09:23:29.978945502 +0000 UTC m=+777.291556650" observedRunningTime="2026-03-14 09:23:30.561671663 +0000 UTC m=+777.874282781" watchObservedRunningTime="2026-03-14 09:23:30.565681134 +0000 UTC m=+777.878292262" Mar 14 09:23:36 crc kubenswrapper[4843]: I0314 09:23:36.732559 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:36 crc kubenswrapper[4843]: I0314 09:23:36.732900 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:37 crc kubenswrapper[4843]: I0314 09:23:37.777579 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zhsnn" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="registry-server" probeResult="failure" output=< Mar 14 09:23:37 crc kubenswrapper[4843]: timeout: failed to connect service ":50051" within 1s Mar 14 09:23:37 crc kubenswrapper[4843]: > Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.034939 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9"] Mar 14 09:23:41 crc kubenswrapper[4843]: E0314 09:23:41.035510 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="extract" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.035527 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="extract" Mar 14 09:23:41 crc kubenswrapper[4843]: E0314 09:23:41.035550 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="pull" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.035558 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="pull" Mar 14 09:23:41 crc kubenswrapper[4843]: E0314 09:23:41.035571 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="util" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.035578 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="util" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.035701 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d916f254-7cc8-48b9-b01e-ba6270584eb3" containerName="extract" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.036219 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.039359 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.039733 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-qccng" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.040058 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.040100 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.040661 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.065127 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9"] Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.216177 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1501c094-2079-46b0-8096-b4a09ee21e8b-webhook-cert\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.216247 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-776gc\" (UniqueName: \"kubernetes.io/projected/1501c094-2079-46b0-8096-b4a09ee21e8b-kube-api-access-776gc\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.216289 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1501c094-2079-46b0-8096-b4a09ee21e8b-apiservice-cert\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.317590 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-776gc\" (UniqueName: \"kubernetes.io/projected/1501c094-2079-46b0-8096-b4a09ee21e8b-kube-api-access-776gc\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.317920 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1501c094-2079-46b0-8096-b4a09ee21e8b-apiservice-cert\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.318094 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1501c094-2079-46b0-8096-b4a09ee21e8b-webhook-cert\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.326122 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1501c094-2079-46b0-8096-b4a09ee21e8b-webhook-cert\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.326155 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1501c094-2079-46b0-8096-b4a09ee21e8b-apiservice-cert\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.335544 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-776gc\" (UniqueName: \"kubernetes.io/projected/1501c094-2079-46b0-8096-b4a09ee21e8b-kube-api-access-776gc\") pod \"metallb-operator-controller-manager-6c5855cb99-gcpx9\" (UID: \"1501c094-2079-46b0-8096-b4a09ee21e8b\") " pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.354368 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.361460 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6"] Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.362372 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.364135 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.364410 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-txrs5" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.364621 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.417160 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6"] Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.520284 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebaf629f-8937-4e21-bedb-0d5992f22edc-webhook-cert\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.520454 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebaf629f-8937-4e21-bedb-0d5992f22edc-apiservice-cert\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.520483 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stpl9\" (UniqueName: \"kubernetes.io/projected/ebaf629f-8937-4e21-bedb-0d5992f22edc-kube-api-access-stpl9\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.623987 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebaf629f-8937-4e21-bedb-0d5992f22edc-apiservice-cert\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.624372 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stpl9\" (UniqueName: \"kubernetes.io/projected/ebaf629f-8937-4e21-bedb-0d5992f22edc-kube-api-access-stpl9\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.624428 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebaf629f-8937-4e21-bedb-0d5992f22edc-webhook-cert\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.628826 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebaf629f-8937-4e21-bedb-0d5992f22edc-apiservice-cert\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.632016 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebaf629f-8937-4e21-bedb-0d5992f22edc-webhook-cert\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.691224 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stpl9\" (UniqueName: \"kubernetes.io/projected/ebaf629f-8937-4e21-bedb-0d5992f22edc-kube-api-access-stpl9\") pod \"metallb-operator-webhook-server-76b978c99b-flxj6\" (UID: \"ebaf629f-8937-4e21-bedb-0d5992f22edc\") " pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.723576 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.785930 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9"] Mar 14 09:23:41 crc kubenswrapper[4843]: I0314 09:23:41.991782 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6"] Mar 14 09:23:41 crc kubenswrapper[4843]: W0314 09:23:41.997684 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebaf629f_8937_4e21_bedb_0d5992f22edc.slice/crio-7abbe7ba35b0ea4480e5a4c44131da2add51fa047adaecb73473710ddc3c28f3 WatchSource:0}: Error finding container 7abbe7ba35b0ea4480e5a4c44131da2add51fa047adaecb73473710ddc3c28f3: Status 404 returned error can't find the container with id 7abbe7ba35b0ea4480e5a4c44131da2add51fa047adaecb73473710ddc3c28f3 Mar 14 09:23:42 crc kubenswrapper[4843]: I0314 09:23:42.629088 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" event={"ID":"ebaf629f-8937-4e21-bedb-0d5992f22edc","Type":"ContainerStarted","Data":"7abbe7ba35b0ea4480e5a4c44131da2add51fa047adaecb73473710ddc3c28f3"} Mar 14 09:23:42 crc kubenswrapper[4843]: I0314 09:23:42.631095 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" event={"ID":"1501c094-2079-46b0-8096-b4a09ee21e8b","Type":"ContainerStarted","Data":"532bab4340436cd59ed0bb803a9f0748fc8cb21f81a70890be759d4b7a65c27a"} Mar 14 09:23:45 crc kubenswrapper[4843]: I0314 09:23:45.387163 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:23:45 crc kubenswrapper[4843]: I0314 09:23:45.387730 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:23:46 crc kubenswrapper[4843]: I0314 09:23:46.773540 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:46 crc kubenswrapper[4843]: I0314 09:23:46.822527 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:48 crc kubenswrapper[4843]: I0314 09:23:48.690673 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" event={"ID":"ebaf629f-8937-4e21-bedb-0d5992f22edc","Type":"ContainerStarted","Data":"812ab1e07f735b555e5bb6db6511e78dddcd0285bb892a7521669a7e1b939dcd"} Mar 14 09:23:48 crc kubenswrapper[4843]: I0314 09:23:48.691026 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:23:48 crc kubenswrapper[4843]: I0314 09:23:48.693647 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" event={"ID":"1501c094-2079-46b0-8096-b4a09ee21e8b","Type":"ContainerStarted","Data":"d9263a01d3bab032dd32cf71b9fb7fcace7cf585b8c1e561511a193b99e77cb8"} Mar 14 09:23:48 crc kubenswrapper[4843]: I0314 09:23:48.693878 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:23:48 crc kubenswrapper[4843]: I0314 09:23:48.730263 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" podStartSLOduration=1.27396273 podStartE2EDuration="7.730229458s" podCreationTimestamp="2026-03-14 09:23:41 +0000 UTC" firstStartedPulling="2026-03-14 09:23:42.000907547 +0000 UTC m=+789.313518685" lastFinishedPulling="2026-03-14 09:23:48.457174285 +0000 UTC m=+795.769785413" observedRunningTime="2026-03-14 09:23:48.70776653 +0000 UTC m=+796.020377658" watchObservedRunningTime="2026-03-14 09:23:48.730229458 +0000 UTC m=+796.042840586" Mar 14 09:23:48 crc kubenswrapper[4843]: I0314 09:23:48.746504 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" podStartSLOduration=1.139981352 podStartE2EDuration="7.746463409s" podCreationTimestamp="2026-03-14 09:23:41 +0000 UTC" firstStartedPulling="2026-03-14 09:23:41.816921245 +0000 UTC m=+789.129532373" lastFinishedPulling="2026-03-14 09:23:48.423403292 +0000 UTC m=+795.736014430" observedRunningTime="2026-03-14 09:23:48.740152769 +0000 UTC m=+796.052763907" watchObservedRunningTime="2026-03-14 09:23:48.746463409 +0000 UTC m=+796.059074537" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.008933 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zhsnn"] Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.009170 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zhsnn" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="registry-server" containerID="cri-o://b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e" gracePeriod=2 Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.372236 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.539719 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-catalog-content\") pod \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.539827 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-utilities\") pod \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.539874 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz7rn\" (UniqueName: \"kubernetes.io/projected/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-kube-api-access-zz7rn\") pod \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\" (UID: \"a2e14476-ee25-47d4-920e-0c8eab9ba7a5\") " Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.540959 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-utilities" (OuterVolumeSpecName: "utilities") pod "a2e14476-ee25-47d4-920e-0c8eab9ba7a5" (UID: "a2e14476-ee25-47d4-920e-0c8eab9ba7a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.554470 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-kube-api-access-zz7rn" (OuterVolumeSpecName: "kube-api-access-zz7rn") pod "a2e14476-ee25-47d4-920e-0c8eab9ba7a5" (UID: "a2e14476-ee25-47d4-920e-0c8eab9ba7a5"). InnerVolumeSpecName "kube-api-access-zz7rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.641162 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.641201 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz7rn\" (UniqueName: \"kubernetes.io/projected/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-kube-api-access-zz7rn\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.673366 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2e14476-ee25-47d4-920e-0c8eab9ba7a5" (UID: "a2e14476-ee25-47d4-920e-0c8eab9ba7a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.701342 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerID="b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e" exitCode=0 Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.701413 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerDied","Data":"b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e"} Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.701492 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zhsnn" event={"ID":"a2e14476-ee25-47d4-920e-0c8eab9ba7a5","Type":"ContainerDied","Data":"73967c6d445069267e3f6f9cf56a1eaa900a5c61b02c43228f1affe82b4d2dac"} Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.701423 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zhsnn" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.701538 4843 scope.go:117] "RemoveContainer" containerID="b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.717055 4843 scope.go:117] "RemoveContainer" containerID="7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.731124 4843 scope.go:117] "RemoveContainer" containerID="ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.742444 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e14476-ee25-47d4-920e-0c8eab9ba7a5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.745833 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zhsnn"] Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.750760 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zhsnn"] Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.772338 4843 scope.go:117] "RemoveContainer" containerID="b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e" Mar 14 09:23:49 crc kubenswrapper[4843]: E0314 09:23:49.772677 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e\": container with ID starting with b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e not found: ID does not exist" containerID="b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.772711 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e"} err="failed to get container status \"b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e\": rpc error: code = NotFound desc = could not find container \"b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e\": container with ID starting with b071e0ddc8dfc23afc449c51ab1a4420d4aa5bb88d270d984b6242b0ba4e098e not found: ID does not exist" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.772736 4843 scope.go:117] "RemoveContainer" containerID="7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789" Mar 14 09:23:49 crc kubenswrapper[4843]: E0314 09:23:49.773109 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789\": container with ID starting with 7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789 not found: ID does not exist" containerID="7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.773137 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789"} err="failed to get container status \"7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789\": rpc error: code = NotFound desc = could not find container \"7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789\": container with ID starting with 7d8010e6631002e71dcebfa1e6cdc72b39b72a67a7065ae37ee63e7f1db37789 not found: ID does not exist" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.773154 4843 scope.go:117] "RemoveContainer" containerID="ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de" Mar 14 09:23:49 crc kubenswrapper[4843]: E0314 09:23:49.773537 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de\": container with ID starting with ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de not found: ID does not exist" containerID="ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de" Mar 14 09:23:49 crc kubenswrapper[4843]: I0314 09:23:49.773561 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de"} err="failed to get container status \"ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de\": rpc error: code = NotFound desc = could not find container \"ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de\": container with ID starting with ff7b83b8c9c481ed021e1bd9443a864f3a02300affb0579b56334692b91c18de not found: ID does not exist" Mar 14 09:23:51 crc kubenswrapper[4843]: I0314 09:23:51.346210 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" path="/var/lib/kubelet/pods/a2e14476-ee25-47d4-920e-0c8eab9ba7a5/volumes" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.130044 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558004-jlkwf"] Mar 14 09:24:00 crc kubenswrapper[4843]: E0314 09:24:00.130907 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="registry-server" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.130926 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="registry-server" Mar 14 09:24:00 crc kubenswrapper[4843]: E0314 09:24:00.130944 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="extract-content" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.130951 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="extract-content" Mar 14 09:24:00 crc kubenswrapper[4843]: E0314 09:24:00.130967 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="extract-utilities" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.130976 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="extract-utilities" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.131103 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e14476-ee25-47d4-920e-0c8eab9ba7a5" containerName="registry-server" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.131711 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.134208 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.134323 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.134390 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.136319 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558004-jlkwf"] Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.170942 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-498kv\" (UniqueName: \"kubernetes.io/projected/8ff20372-3c55-4465-8ddf-8467007709a3-kube-api-access-498kv\") pod \"auto-csr-approver-29558004-jlkwf\" (UID: \"8ff20372-3c55-4465-8ddf-8467007709a3\") " pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.271830 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-498kv\" (UniqueName: \"kubernetes.io/projected/8ff20372-3c55-4465-8ddf-8467007709a3-kube-api-access-498kv\") pod \"auto-csr-approver-29558004-jlkwf\" (UID: \"8ff20372-3c55-4465-8ddf-8467007709a3\") " pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.300523 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-498kv\" (UniqueName: \"kubernetes.io/projected/8ff20372-3c55-4465-8ddf-8467007709a3-kube-api-access-498kv\") pod \"auto-csr-approver-29558004-jlkwf\" (UID: \"8ff20372-3c55-4465-8ddf-8467007709a3\") " pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.471728 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.708534 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558004-jlkwf"] Mar 14 09:24:00 crc kubenswrapper[4843]: I0314 09:24:00.767708 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" event={"ID":"8ff20372-3c55-4465-8ddf-8467007709a3","Type":"ContainerStarted","Data":"2740d7fe9828e4e6afdf414df29f42c1b456a2b71136193571dcf1dfcb94896f"} Mar 14 09:24:01 crc kubenswrapper[4843]: I0314 09:24:01.728602 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-76b978c99b-flxj6" Mar 14 09:24:02 crc kubenswrapper[4843]: I0314 09:24:02.793036 4843 generic.go:334] "Generic (PLEG): container finished" podID="8ff20372-3c55-4465-8ddf-8467007709a3" containerID="3ff8c1547f4171613d9750fb339269d2ecf0f9df9397830da34db78298934b49" exitCode=0 Mar 14 09:24:02 crc kubenswrapper[4843]: I0314 09:24:02.793076 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" event={"ID":"8ff20372-3c55-4465-8ddf-8467007709a3","Type":"ContainerDied","Data":"3ff8c1547f4171613d9750fb339269d2ecf0f9df9397830da34db78298934b49"} Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.108829 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.128681 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-498kv\" (UniqueName: \"kubernetes.io/projected/8ff20372-3c55-4465-8ddf-8467007709a3-kube-api-access-498kv\") pod \"8ff20372-3c55-4465-8ddf-8467007709a3\" (UID: \"8ff20372-3c55-4465-8ddf-8467007709a3\") " Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.134689 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff20372-3c55-4465-8ddf-8467007709a3-kube-api-access-498kv" (OuterVolumeSpecName: "kube-api-access-498kv") pod "8ff20372-3c55-4465-8ddf-8467007709a3" (UID: "8ff20372-3c55-4465-8ddf-8467007709a3"). InnerVolumeSpecName "kube-api-access-498kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.234948 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-498kv\" (UniqueName: \"kubernetes.io/projected/8ff20372-3c55-4465-8ddf-8467007709a3-kube-api-access-498kv\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.807677 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" event={"ID":"8ff20372-3c55-4465-8ddf-8467007709a3","Type":"ContainerDied","Data":"2740d7fe9828e4e6afdf414df29f42c1b456a2b71136193571dcf1dfcb94896f"} Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.807718 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2740d7fe9828e4e6afdf414df29f42c1b456a2b71136193571dcf1dfcb94896f" Mar 14 09:24:04 crc kubenswrapper[4843]: I0314 09:24:04.807766 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558004-jlkwf" Mar 14 09:24:05 crc kubenswrapper[4843]: I0314 09:24:05.169435 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557998-pjtf2"] Mar 14 09:24:05 crc kubenswrapper[4843]: I0314 09:24:05.175703 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557998-pjtf2"] Mar 14 09:24:05 crc kubenswrapper[4843]: I0314 09:24:05.345877 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="494e5ab2-0422-4ac9-b963-15303eb3dcea" path="/var/lib/kubelet/pods/494e5ab2-0422-4ac9-b963-15303eb3dcea/volumes" Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.387260 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.387597 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.387651 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.388203 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0419057bafaf42ee3e8fe2dcca9e3ad2340132e208a0d8b9f235557b85f11481"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.388256 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://0419057bafaf42ee3e8fe2dcca9e3ad2340132e208a0d8b9f235557b85f11481" gracePeriod=600 Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.881039 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="0419057bafaf42ee3e8fe2dcca9e3ad2340132e208a0d8b9f235557b85f11481" exitCode=0 Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.881094 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"0419057bafaf42ee3e8fe2dcca9e3ad2340132e208a0d8b9f235557b85f11481"} Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.881380 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"144fcfbfce69a65758af21b0b25db69c4308200478b4f5a2c753f3696a587b2b"} Mar 14 09:24:15 crc kubenswrapper[4843]: I0314 09:24:15.881403 4843 scope.go:117] "RemoveContainer" containerID="11847d99c55208b09be55f12335af74a9e648d402160cb65deed3a9467da14ac" Mar 14 09:24:21 crc kubenswrapper[4843]: I0314 09:24:21.357293 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6c5855cb99-gcpx9" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.143710 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2llz5"] Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.143984 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff20372-3c55-4465-8ddf-8467007709a3" containerName="oc" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.144004 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff20372-3c55-4465-8ddf-8467007709a3" containerName="oc" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.144165 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff20372-3c55-4465-8ddf-8467007709a3" containerName="oc" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.146586 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.149107 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.149126 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.149734 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ltv4t" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.150760 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f"] Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.151581 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.155731 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.168920 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f"] Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219320 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-frr-sockets\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219357 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/63a81c2e-7222-4852-b557-f6a24194d46f-frr-startup\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219374 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-reloader\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219389 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-metrics\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219427 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv558\" (UniqueName: \"kubernetes.io/projected/63a81c2e-7222-4852-b557-f6a24194d46f-kube-api-access-gv558\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219491 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78jgq\" (UniqueName: \"kubernetes.io/projected/2022ba85-6e88-485c-ba26-7f16fc03c257-kube-api-access-78jgq\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219549 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/63a81c2e-7222-4852-b557-f6a24194d46f-metrics-certs\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219585 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2022ba85-6e88-485c-ba26-7f16fc03c257-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.219638 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-frr-conf\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.239637 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bv84h"] Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.240766 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.242344 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.242940 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.243115 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-vfftb" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.245265 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.253825 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-8v9j7"] Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.254752 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.264804 4843 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.271829 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-8v9j7"] Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320565 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-frr-sockets\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320606 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/63a81c2e-7222-4852-b557-f6a24194d46f-frr-startup\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320622 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-reloader\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320638 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-metrics\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320661 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv558\" (UniqueName: \"kubernetes.io/projected/63a81c2e-7222-4852-b557-f6a24194d46f-kube-api-access-gv558\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320685 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78jgq\" (UniqueName: \"kubernetes.io/projected/2022ba85-6e88-485c-ba26-7f16fc03c257-kube-api-access-78jgq\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320699 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/63a81c2e-7222-4852-b557-f6a24194d46f-metrics-certs\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320723 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2022ba85-6e88-485c-ba26-7f16fc03c257-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.320737 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-frr-conf\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.321071 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-frr-conf\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.321257 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-frr-sockets\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.321911 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/63a81c2e-7222-4852-b557-f6a24194d46f-frr-startup\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.322109 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-reloader\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.322334 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/63a81c2e-7222-4852-b557-f6a24194d46f-metrics\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.323648 4843 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.323711 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2022ba85-6e88-485c-ba26-7f16fc03c257-cert podName:2022ba85-6e88-485c-ba26-7f16fc03c257 nodeName:}" failed. No retries permitted until 2026-03-14 09:24:22.823698442 +0000 UTC m=+830.136309570 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2022ba85-6e88-485c-ba26-7f16fc03c257-cert") pod "frr-k8s-webhook-server-bcc4b6f68-nbb5f" (UID: "2022ba85-6e88-485c-ba26-7f16fc03c257") : secret "frr-k8s-webhook-server-cert" not found Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.328814 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/63a81c2e-7222-4852-b557-f6a24194d46f-metrics-certs\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.341935 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78jgq\" (UniqueName: \"kubernetes.io/projected/2022ba85-6e88-485c-ba26-7f16fc03c257-kube-api-access-78jgq\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.341965 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv558\" (UniqueName: \"kubernetes.io/projected/63a81c2e-7222-4852-b557-f6a24194d46f-kube-api-access-gv558\") pod \"frr-k8s-2llz5\" (UID: \"63a81c2e-7222-4852-b557-f6a24194d46f\") " pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422178 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-metrics-certs\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422243 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-memberlist\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422306 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3444025-ba10-41ec-94cd-0ce213c73311-metallb-excludel2\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422342 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vgm7\" (UniqueName: \"kubernetes.io/projected/d3444025-ba10-41ec-94cd-0ce213c73311-kube-api-access-2vgm7\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422384 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-cert\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422411 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96bvq\" (UniqueName: \"kubernetes.io/projected/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-kube-api-access-96bvq\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.422450 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-metrics-certs\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.467390 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523665 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-metrics-certs\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523718 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-memberlist\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523760 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3444025-ba10-41ec-94cd-0ce213c73311-metallb-excludel2\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523791 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vgm7\" (UniqueName: \"kubernetes.io/projected/d3444025-ba10-41ec-94cd-0ce213c73311-kube-api-access-2vgm7\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523811 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-cert\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523832 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96bvq\" (UniqueName: \"kubernetes.io/projected/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-kube-api-access-96bvq\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.523855 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-metrics-certs\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.524420 4843 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.524502 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-memberlist podName:d3444025-ba10-41ec-94cd-0ce213c73311 nodeName:}" failed. No retries permitted until 2026-03-14 09:24:23.024477277 +0000 UTC m=+830.337088615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-memberlist") pod "speaker-bv84h" (UID: "d3444025-ba10-41ec-94cd-0ce213c73311") : secret "metallb-memberlist" not found Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.524819 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3444025-ba10-41ec-94cd-0ce213c73311-metallb-excludel2\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.526496 4843 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Mar 14 09:24:22 crc kubenswrapper[4843]: E0314 09:24:22.526528 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-metrics-certs podName:d3444025-ba10-41ec-94cd-0ce213c73311 nodeName:}" failed. No retries permitted until 2026-03-14 09:24:23.026517099 +0000 UTC m=+830.339128227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-metrics-certs") pod "speaker-bv84h" (UID: "d3444025-ba10-41ec-94cd-0ce213c73311") : secret "speaker-certs-secret" not found Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.529674 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-metrics-certs\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.530532 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-cert\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.540902 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96bvq\" (UniqueName: \"kubernetes.io/projected/64b1131f-66f6-4bdf-b60e-f5d8f8680eeb-kube-api-access-96bvq\") pod \"controller-7bb4cc7c98-8v9j7\" (UID: \"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb\") " pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.564300 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vgm7\" (UniqueName: \"kubernetes.io/projected/d3444025-ba10-41ec-94cd-0ce213c73311-kube-api-access-2vgm7\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.570538 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.781762 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-8v9j7"] Mar 14 09:24:22 crc kubenswrapper[4843]: W0314 09:24:22.785820 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64b1131f_66f6_4bdf_b60e_f5d8f8680eeb.slice/crio-02588980f1988f43b04cf1e1ac202c43e60f9ea2f3146eec4fb8544ebde9b100 WatchSource:0}: Error finding container 02588980f1988f43b04cf1e1ac202c43e60f9ea2f3146eec4fb8544ebde9b100: Status 404 returned error can't find the container with id 02588980f1988f43b04cf1e1ac202c43e60f9ea2f3146eec4fb8544ebde9b100 Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.829005 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2022ba85-6e88-485c-ba26-7f16fc03c257-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.834415 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2022ba85-6e88-485c-ba26-7f16fc03c257-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nbb5f\" (UID: \"2022ba85-6e88-485c-ba26-7f16fc03c257\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.937027 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-8v9j7" event={"ID":"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb","Type":"ContainerStarted","Data":"20ecae1e8fc1cac9255f7711da69d44c941dc34329caa4efeed7c838f14e031e"} Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.937078 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-8v9j7" event={"ID":"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb","Type":"ContainerStarted","Data":"02588980f1988f43b04cf1e1ac202c43e60f9ea2f3146eec4fb8544ebde9b100"} Mar 14 09:24:22 crc kubenswrapper[4843]: I0314 09:24:22.938121 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"fd008d54b5c9ce448eafb29acced479998928b64de38b315393ecab506aee1dc"} Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.033715 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-metrics-certs\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.033757 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-memberlist\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.037284 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-metrics-certs\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.037689 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3444025-ba10-41ec-94cd-0ce213c73311-memberlist\") pod \"speaker-bv84h\" (UID: \"d3444025-ba10-41ec-94cd-0ce213c73311\") " pod="metallb-system/speaker-bv84h" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.076523 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.154940 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bv84h" Mar 14 09:24:23 crc kubenswrapper[4843]: W0314 09:24:23.212554 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3444025_ba10_41ec_94cd_0ce213c73311.slice/crio-852834f1d9bf0c7cbc1e314dd96cc20e74987a47ec12032db7666108d7d8c07c WatchSource:0}: Error finding container 852834f1d9bf0c7cbc1e314dd96cc20e74987a47ec12032db7666108d7d8c07c: Status 404 returned error can't find the container with id 852834f1d9bf0c7cbc1e314dd96cc20e74987a47ec12032db7666108d7d8c07c Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.544347 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f"] Mar 14 09:24:23 crc kubenswrapper[4843]: W0314 09:24:23.549909 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2022ba85_6e88_485c_ba26_7f16fc03c257.slice/crio-e463b9906a5dd86c757e92a534f25e68cd7fe411acbccc9ba2c6c11ebde1a90a WatchSource:0}: Error finding container e463b9906a5dd86c757e92a534f25e68cd7fe411acbccc9ba2c6c11ebde1a90a: Status 404 returned error can't find the container with id e463b9906a5dd86c757e92a534f25e68cd7fe411acbccc9ba2c6c11ebde1a90a Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.947595 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" event={"ID":"2022ba85-6e88-485c-ba26-7f16fc03c257","Type":"ContainerStarted","Data":"e463b9906a5dd86c757e92a534f25e68cd7fe411acbccc9ba2c6c11ebde1a90a"} Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.949588 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bv84h" event={"ID":"d3444025-ba10-41ec-94cd-0ce213c73311","Type":"ContainerStarted","Data":"0f7b73e62f158b5cd613d009978afaa6d6b8e24659e14f7e0b5e01d89e0ffaca"} Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.949640 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bv84h" event={"ID":"d3444025-ba10-41ec-94cd-0ce213c73311","Type":"ContainerStarted","Data":"ad47085f713051382506cb374f6e0f61fd8b36bd3d2745d62e98213d76049e2d"} Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.949655 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bv84h" event={"ID":"d3444025-ba10-41ec-94cd-0ce213c73311","Type":"ContainerStarted","Data":"852834f1d9bf0c7cbc1e314dd96cc20e74987a47ec12032db7666108d7d8c07c"} Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.949825 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bv84h" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.951589 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-8v9j7" event={"ID":"64b1131f-66f6-4bdf-b60e-f5d8f8680eeb","Type":"ContainerStarted","Data":"81c7ac05351ee9cfc6548c21cff12badbbbbe5a1dbbad80daa8a203dfd60cb3c"} Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.951802 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.967704 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bv84h" podStartSLOduration=1.9676842799999998 podStartE2EDuration="1.96768428s" podCreationTimestamp="2026-03-14 09:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:24:23.96409443 +0000 UTC m=+831.276705568" watchObservedRunningTime="2026-03-14 09:24:23.96768428 +0000 UTC m=+831.280295408" Mar 14 09:24:23 crc kubenswrapper[4843]: I0314 09:24:23.981992 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-8v9j7" podStartSLOduration=1.9819706510000001 podStartE2EDuration="1.981970651s" podCreationTimestamp="2026-03-14 09:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:24:23.980447703 +0000 UTC m=+831.293058861" watchObservedRunningTime="2026-03-14 09:24:23.981970651 +0000 UTC m=+831.294581789" Mar 14 09:24:31 crc kubenswrapper[4843]: I0314 09:24:31.003918 4843 generic.go:334] "Generic (PLEG): container finished" podID="63a81c2e-7222-4852-b557-f6a24194d46f" containerID="57e49b92b83c74d9becd6847d012100372b59ef5d60c67e966f07109224baec2" exitCode=0 Mar 14 09:24:31 crc kubenswrapper[4843]: I0314 09:24:31.004172 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerDied","Data":"57e49b92b83c74d9becd6847d012100372b59ef5d60c67e966f07109224baec2"} Mar 14 09:24:31 crc kubenswrapper[4843]: I0314 09:24:31.015352 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" event={"ID":"2022ba85-6e88-485c-ba26-7f16fc03c257","Type":"ContainerStarted","Data":"56209d7cbbd24e5816ba20e6e4a1fcb751f69246c72223400e245d98d9b456c6"} Mar 14 09:24:31 crc kubenswrapper[4843]: I0314 09:24:31.015728 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:31 crc kubenswrapper[4843]: I0314 09:24:31.053702 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" podStartSLOduration=2.459574445 podStartE2EDuration="9.053679508s" podCreationTimestamp="2026-03-14 09:24:22 +0000 UTC" firstStartedPulling="2026-03-14 09:24:23.554951077 +0000 UTC m=+830.867562205" lastFinishedPulling="2026-03-14 09:24:30.14905613 +0000 UTC m=+837.461667268" observedRunningTime="2026-03-14 09:24:31.05100035 +0000 UTC m=+838.363611488" watchObservedRunningTime="2026-03-14 09:24:31.053679508 +0000 UTC m=+838.366290646" Mar 14 09:24:32 crc kubenswrapper[4843]: I0314 09:24:32.021630 4843 generic.go:334] "Generic (PLEG): container finished" podID="63a81c2e-7222-4852-b557-f6a24194d46f" containerID="cc8d6a3dbcfee78042acdfbf1e3f1e2059616620581eb12622543b9446ef1d00" exitCode=0 Mar 14 09:24:32 crc kubenswrapper[4843]: I0314 09:24:32.021662 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerDied","Data":"cc8d6a3dbcfee78042acdfbf1e3f1e2059616620581eb12622543b9446ef1d00"} Mar 14 09:24:32 crc kubenswrapper[4843]: I0314 09:24:32.574054 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-8v9j7" Mar 14 09:24:33 crc kubenswrapper[4843]: I0314 09:24:33.158881 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bv84h" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.036510 4843 generic.go:334] "Generic (PLEG): container finished" podID="63a81c2e-7222-4852-b557-f6a24194d46f" containerID="29be21c9e3d0da47d102686ccf4e8b2dd5856e48790447bd0ba99804e0259acf" exitCode=0 Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.036604 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerDied","Data":"29be21c9e3d0da47d102686ccf4e8b2dd5856e48790447bd0ba99804e0259acf"} Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.663857 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx"] Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.665532 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.670057 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.673826 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx"] Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.693966 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz85m\" (UniqueName: \"kubernetes.io/projected/bd5a3603-199d-45f6-80c7-afb96e29068f-kube-api-access-xz85m\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.694015 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.694078 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.795731 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.795832 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz85m\" (UniqueName: \"kubernetes.io/projected/bd5a3603-199d-45f6-80c7-afb96e29068f-kube-api-access-xz85m\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.795860 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.796652 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.799808 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.814044 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz85m\" (UniqueName: \"kubernetes.io/projected/bd5a3603-199d-45f6-80c7-afb96e29068f-kube-api-access-xz85m\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:34 crc kubenswrapper[4843]: I0314 09:24:34.982518 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:35 crc kubenswrapper[4843]: I0314 09:24:35.047123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"6298bad130f1a5ed2046657de9e184755e77b7eeee5e70393fe9cb00bcea20fe"} Mar 14 09:24:35 crc kubenswrapper[4843]: I0314 09:24:35.047167 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"bf1e1417524617bc03a6cbad91fa02b7720c3b2041155eb62527a2e0d82280c8"} Mar 14 09:24:35 crc kubenswrapper[4843]: I0314 09:24:35.047177 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"89f83c49a92b7e6bfa22caab26ccc61a0b0a9fcda1f50238713a17593e081f76"} Mar 14 09:24:35 crc kubenswrapper[4843]: I0314 09:24:35.047185 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"5638849d947d212bc1c5b5f8801a98f629ef2f06f539d669a2d55d68571b4ca5"} Mar 14 09:24:35 crc kubenswrapper[4843]: I0314 09:24:35.047195 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"cb4dd738ac33fa7017641990b772f0ef3f5a725d9c1a8126740ec9a31bc403b3"} Mar 14 09:24:35 crc kubenswrapper[4843]: I0314 09:24:35.427353 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx"] Mar 14 09:24:36 crc kubenswrapper[4843]: I0314 09:24:36.054622 4843 generic.go:334] "Generic (PLEG): container finished" podID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerID="251402a0f5983f08e74ed2210a6415286219e276baa08285d27c481e948ce4a1" exitCode=0 Mar 14 09:24:36 crc kubenswrapper[4843]: I0314 09:24:36.054956 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" event={"ID":"bd5a3603-199d-45f6-80c7-afb96e29068f","Type":"ContainerDied","Data":"251402a0f5983f08e74ed2210a6415286219e276baa08285d27c481e948ce4a1"} Mar 14 09:24:36 crc kubenswrapper[4843]: I0314 09:24:36.054987 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" event={"ID":"bd5a3603-199d-45f6-80c7-afb96e29068f","Type":"ContainerStarted","Data":"5a6f077a2b78e746929829cce013cb9aeb845504d42819b707a796f9b89c62ed"} Mar 14 09:24:36 crc kubenswrapper[4843]: I0314 09:24:36.060480 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2llz5" event={"ID":"63a81c2e-7222-4852-b557-f6a24194d46f","Type":"ContainerStarted","Data":"0d3b9ed2fa95c9a9d6e76df4567acdc4b9f192a15db2caa1afa0b1513c950d86"} Mar 14 09:24:36 crc kubenswrapper[4843]: I0314 09:24:36.060747 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:36 crc kubenswrapper[4843]: I0314 09:24:36.098050 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2llz5" podStartSLOduration=6.632082983 podStartE2EDuration="14.098031105s" podCreationTimestamp="2026-03-14 09:24:22 +0000 UTC" firstStartedPulling="2026-03-14 09:24:22.658363512 +0000 UTC m=+829.970974640" lastFinishedPulling="2026-03-14 09:24:30.124311634 +0000 UTC m=+837.436922762" observedRunningTime="2026-03-14 09:24:36.096968998 +0000 UTC m=+843.409580136" watchObservedRunningTime="2026-03-14 09:24:36.098031105 +0000 UTC m=+843.410642223" Mar 14 09:24:37 crc kubenswrapper[4843]: I0314 09:24:37.468224 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:37 crc kubenswrapper[4843]: I0314 09:24:37.526588 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:40 crc kubenswrapper[4843]: I0314 09:24:40.088682 4843 generic.go:334] "Generic (PLEG): container finished" podID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerID="d0f9895decb92510940d725d6b7c878d0dde959ae1c8c8278aeb86ff3f8a47bc" exitCode=0 Mar 14 09:24:40 crc kubenswrapper[4843]: I0314 09:24:40.088830 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" event={"ID":"bd5a3603-199d-45f6-80c7-afb96e29068f","Type":"ContainerDied","Data":"d0f9895decb92510940d725d6b7c878d0dde959ae1c8c8278aeb86ff3f8a47bc"} Mar 14 09:24:40 crc kubenswrapper[4843]: I0314 09:24:40.435941 4843 scope.go:117] "RemoveContainer" containerID="9a649913f4f7f5d84c93acf1d9dafd9a61e47d9b2247f5e7c364c17ed852a9e3" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.099229 4843 generic.go:334] "Generic (PLEG): container finished" podID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerID="c4fc1a71a5d3b8afbd24805f0697275197a3501c9f065d061097cdf82b895900" exitCode=0 Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.099565 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" event={"ID":"bd5a3603-199d-45f6-80c7-afb96e29068f","Type":"ContainerDied","Data":"c4fc1a71a5d3b8afbd24805f0697275197a3501c9f065d061097cdf82b895900"} Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.189943 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j9wqn"] Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.191629 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.205535 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9wqn"] Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.293654 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-utilities\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.293780 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtflq\" (UniqueName: \"kubernetes.io/projected/61b8979b-1447-4cb8-b87e-348f713d002b-kube-api-access-vtflq\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.293833 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-catalog-content\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.395008 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtflq\" (UniqueName: \"kubernetes.io/projected/61b8979b-1447-4cb8-b87e-348f713d002b-kube-api-access-vtflq\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.395080 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-catalog-content\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.395134 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-utilities\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.395629 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-utilities\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.395825 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-catalog-content\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.427151 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtflq\" (UniqueName: \"kubernetes.io/projected/61b8979b-1447-4cb8-b87e-348f713d002b-kube-api-access-vtflq\") pod \"community-operators-j9wqn\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:41 crc kubenswrapper[4843]: I0314 09:24:41.515809 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.042617 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9wqn"] Mar 14 09:24:42 crc kubenswrapper[4843]: W0314 09:24:42.054482 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61b8979b_1447_4cb8_b87e_348f713d002b.slice/crio-6d0391d4caa79061a6c3940e4d2839f6049f66343f2e80d044973c76e46c2efa WatchSource:0}: Error finding container 6d0391d4caa79061a6c3940e4d2839f6049f66343f2e80d044973c76e46c2efa: Status 404 returned error can't find the container with id 6d0391d4caa79061a6c3940e4d2839f6049f66343f2e80d044973c76e46c2efa Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.106641 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wqn" event={"ID":"61b8979b-1447-4cb8-b87e-348f713d002b","Type":"ContainerStarted","Data":"6d0391d4caa79061a6c3940e4d2839f6049f66343f2e80d044973c76e46c2efa"} Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.396667 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.522749 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz85m\" (UniqueName: \"kubernetes.io/projected/bd5a3603-199d-45f6-80c7-afb96e29068f-kube-api-access-xz85m\") pod \"bd5a3603-199d-45f6-80c7-afb96e29068f\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.522881 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-bundle\") pod \"bd5a3603-199d-45f6-80c7-afb96e29068f\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.522928 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-util\") pod \"bd5a3603-199d-45f6-80c7-afb96e29068f\" (UID: \"bd5a3603-199d-45f6-80c7-afb96e29068f\") " Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.524732 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-bundle" (OuterVolumeSpecName: "bundle") pod "bd5a3603-199d-45f6-80c7-afb96e29068f" (UID: "bd5a3603-199d-45f6-80c7-afb96e29068f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.530573 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd5a3603-199d-45f6-80c7-afb96e29068f-kube-api-access-xz85m" (OuterVolumeSpecName: "kube-api-access-xz85m") pod "bd5a3603-199d-45f6-80c7-afb96e29068f" (UID: "bd5a3603-199d-45f6-80c7-afb96e29068f"). InnerVolumeSpecName "kube-api-access-xz85m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.540592 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-util" (OuterVolumeSpecName: "util") pod "bd5a3603-199d-45f6-80c7-afb96e29068f" (UID: "bd5a3603-199d-45f6-80c7-afb96e29068f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.625661 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz85m\" (UniqueName: \"kubernetes.io/projected/bd5a3603-199d-45f6-80c7-afb96e29068f-kube-api-access-xz85m\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.625742 4843 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:42 crc kubenswrapper[4843]: I0314 09:24:42.625765 4843 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd5a3603-199d-45f6-80c7-afb96e29068f-util\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:43 crc kubenswrapper[4843]: I0314 09:24:43.081474 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nbb5f" Mar 14 09:24:43 crc kubenswrapper[4843]: I0314 09:24:43.127444 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" event={"ID":"bd5a3603-199d-45f6-80c7-afb96e29068f","Type":"ContainerDied","Data":"5a6f077a2b78e746929829cce013cb9aeb845504d42819b707a796f9b89c62ed"} Mar 14 09:24:43 crc kubenswrapper[4843]: I0314 09:24:43.127506 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a6f077a2b78e746929829cce013cb9aeb845504d42819b707a796f9b89c62ed" Mar 14 09:24:43 crc kubenswrapper[4843]: I0314 09:24:43.127602 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx" Mar 14 09:24:43 crc kubenswrapper[4843]: I0314 09:24:43.129515 4843 generic.go:334] "Generic (PLEG): container finished" podID="61b8979b-1447-4cb8-b87e-348f713d002b" containerID="69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b" exitCode=0 Mar 14 09:24:43 crc kubenswrapper[4843]: I0314 09:24:43.129558 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wqn" event={"ID":"61b8979b-1447-4cb8-b87e-348f713d002b","Type":"ContainerDied","Data":"69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b"} Mar 14 09:24:45 crc kubenswrapper[4843]: I0314 09:24:45.144487 4843 generic.go:334] "Generic (PLEG): container finished" podID="61b8979b-1447-4cb8-b87e-348f713d002b" containerID="5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55" exitCode=0 Mar 14 09:24:45 crc kubenswrapper[4843]: I0314 09:24:45.144607 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wqn" event={"ID":"61b8979b-1447-4cb8-b87e-348f713d002b","Type":"ContainerDied","Data":"5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55"} Mar 14 09:24:46 crc kubenswrapper[4843]: I0314 09:24:46.164670 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wqn" event={"ID":"61b8979b-1447-4cb8-b87e-348f713d002b","Type":"ContainerStarted","Data":"6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d"} Mar 14 09:24:47 crc kubenswrapper[4843]: I0314 09:24:47.189889 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j9wqn" podStartSLOduration=3.341659754 podStartE2EDuration="6.189867645s" podCreationTimestamp="2026-03-14 09:24:41 +0000 UTC" firstStartedPulling="2026-03-14 09:24:43.1312027 +0000 UTC m=+850.443813828" lastFinishedPulling="2026-03-14 09:24:45.979410591 +0000 UTC m=+853.292021719" observedRunningTime="2026-03-14 09:24:47.183690548 +0000 UTC m=+854.496301676" watchObservedRunningTime="2026-03-14 09:24:47.189867645 +0000 UTC m=+854.502478773" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.583574 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd"] Mar 14 09:24:48 crc kubenswrapper[4843]: E0314 09:24:48.584102 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="extract" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.584118 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="extract" Mar 14 09:24:48 crc kubenswrapper[4843]: E0314 09:24:48.584134 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="util" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.584141 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="util" Mar 14 09:24:48 crc kubenswrapper[4843]: E0314 09:24:48.584156 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="pull" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.584163 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="pull" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.584319 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd5a3603-199d-45f6-80c7-afb96e29068f" containerName="extract" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.584825 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.590008 4843 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-sz2fr" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.590458 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.590662 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.601628 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd"] Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.710635 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpv5j\" (UniqueName: \"kubernetes.io/projected/6b107285-e490-48f4-a158-308c62dd492e-kube-api-access-zpv5j\") pod \"cert-manager-operator-controller-manager-66c8bdd694-mj4wd\" (UID: \"6b107285-e490-48f4-a158-308c62dd492e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.711003 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6b107285-e490-48f4-a158-308c62dd492e-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-mj4wd\" (UID: \"6b107285-e490-48f4-a158-308c62dd492e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.812482 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6b107285-e490-48f4-a158-308c62dd492e-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-mj4wd\" (UID: \"6b107285-e490-48f4-a158-308c62dd492e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.812601 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpv5j\" (UniqueName: \"kubernetes.io/projected/6b107285-e490-48f4-a158-308c62dd492e-kube-api-access-zpv5j\") pod \"cert-manager-operator-controller-manager-66c8bdd694-mj4wd\" (UID: \"6b107285-e490-48f4-a158-308c62dd492e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.813405 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6b107285-e490-48f4-a158-308c62dd492e-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-mj4wd\" (UID: \"6b107285-e490-48f4-a158-308c62dd492e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.840117 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpv5j\" (UniqueName: \"kubernetes.io/projected/6b107285-e490-48f4-a158-308c62dd492e-kube-api-access-zpv5j\") pod \"cert-manager-operator-controller-manager-66c8bdd694-mj4wd\" (UID: \"6b107285-e490-48f4-a158-308c62dd492e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:48 crc kubenswrapper[4843]: I0314 09:24:48.949938 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" Mar 14 09:24:49 crc kubenswrapper[4843]: I0314 09:24:49.179933 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd"] Mar 14 09:24:50 crc kubenswrapper[4843]: I0314 09:24:50.190261 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" event={"ID":"6b107285-e490-48f4-a158-308c62dd492e","Type":"ContainerStarted","Data":"82e19d5e2cd694a799ed28e6324c79d55001e9050d136be5c19e9b39a087585b"} Mar 14 09:24:51 crc kubenswrapper[4843]: I0314 09:24:51.516615 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:51 crc kubenswrapper[4843]: I0314 09:24:51.516665 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:51 crc kubenswrapper[4843]: I0314 09:24:51.560055 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:52 crc kubenswrapper[4843]: I0314 09:24:52.255505 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:52 crc kubenswrapper[4843]: I0314 09:24:52.470603 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2llz5" Mar 14 09:24:52 crc kubenswrapper[4843]: I0314 09:24:52.781744 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9wqn"] Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.216387 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" event={"ID":"6b107285-e490-48f4-a158-308c62dd492e","Type":"ContainerStarted","Data":"a090358a29d5424a0d43bf14d77132604ba775b9710fa6dd2b1ed1cfaf5d814d"} Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.219002 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j9wqn" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="registry-server" containerID="cri-o://6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d" gracePeriod=2 Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.248646 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-mj4wd" podStartSLOduration=2.34591239 podStartE2EDuration="6.248619518s" podCreationTimestamp="2026-03-14 09:24:48 +0000 UTC" firstStartedPulling="2026-03-14 09:24:49.199034383 +0000 UTC m=+856.511645521" lastFinishedPulling="2026-03-14 09:24:53.101741521 +0000 UTC m=+860.414352649" observedRunningTime="2026-03-14 09:24:54.244578502 +0000 UTC m=+861.557189690" watchObservedRunningTime="2026-03-14 09:24:54.248619518 +0000 UTC m=+861.561230646" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.727247 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.807403 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-catalog-content\") pod \"61b8979b-1447-4cb8-b87e-348f713d002b\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.807704 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtflq\" (UniqueName: \"kubernetes.io/projected/61b8979b-1447-4cb8-b87e-348f713d002b-kube-api-access-vtflq\") pod \"61b8979b-1447-4cb8-b87e-348f713d002b\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.807739 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-utilities\") pod \"61b8979b-1447-4cb8-b87e-348f713d002b\" (UID: \"61b8979b-1447-4cb8-b87e-348f713d002b\") " Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.808500 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-utilities" (OuterVolumeSpecName: "utilities") pod "61b8979b-1447-4cb8-b87e-348f713d002b" (UID: "61b8979b-1447-4cb8-b87e-348f713d002b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.813384 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61b8979b-1447-4cb8-b87e-348f713d002b-kube-api-access-vtflq" (OuterVolumeSpecName: "kube-api-access-vtflq") pod "61b8979b-1447-4cb8-b87e-348f713d002b" (UID: "61b8979b-1447-4cb8-b87e-348f713d002b"). InnerVolumeSpecName "kube-api-access-vtflq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.869260 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61b8979b-1447-4cb8-b87e-348f713d002b" (UID: "61b8979b-1447-4cb8-b87e-348f713d002b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.909021 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtflq\" (UniqueName: \"kubernetes.io/projected/61b8979b-1447-4cb8-b87e-348f713d002b-kube-api-access-vtflq\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.909062 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:54 crc kubenswrapper[4843]: I0314 09:24:54.909074 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b8979b-1447-4cb8-b87e-348f713d002b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.224438 4843 generic.go:334] "Generic (PLEG): container finished" podID="61b8979b-1447-4cb8-b87e-348f713d002b" containerID="6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d" exitCode=0 Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.224497 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wqn" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.224526 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wqn" event={"ID":"61b8979b-1447-4cb8-b87e-348f713d002b","Type":"ContainerDied","Data":"6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d"} Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.224588 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wqn" event={"ID":"61b8979b-1447-4cb8-b87e-348f713d002b","Type":"ContainerDied","Data":"6d0391d4caa79061a6c3940e4d2839f6049f66343f2e80d044973c76e46c2efa"} Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.224612 4843 scope.go:117] "RemoveContainer" containerID="6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.247982 4843 scope.go:117] "RemoveContainer" containerID="5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.253310 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9wqn"] Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.257375 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j9wqn"] Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.278530 4843 scope.go:117] "RemoveContainer" containerID="69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.294505 4843 scope.go:117] "RemoveContainer" containerID="6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d" Mar 14 09:24:55 crc kubenswrapper[4843]: E0314 09:24:55.295105 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d\": container with ID starting with 6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d not found: ID does not exist" containerID="6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.295151 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d"} err="failed to get container status \"6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d\": rpc error: code = NotFound desc = could not find container \"6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d\": container with ID starting with 6c616e72f17e3cfaf64179d00c1eea030299e2883f3ccd575cd3e9d967ba027d not found: ID does not exist" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.295182 4843 scope.go:117] "RemoveContainer" containerID="5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55" Mar 14 09:24:55 crc kubenswrapper[4843]: E0314 09:24:55.295520 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55\": container with ID starting with 5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55 not found: ID does not exist" containerID="5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.295548 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55"} err="failed to get container status \"5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55\": rpc error: code = NotFound desc = could not find container \"5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55\": container with ID starting with 5c1bd72beb2088624a0205c462606d2f77aa9591c56421704e1c1c75e9e19a55 not found: ID does not exist" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.295569 4843 scope.go:117] "RemoveContainer" containerID="69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b" Mar 14 09:24:55 crc kubenswrapper[4843]: E0314 09:24:55.295844 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b\": container with ID starting with 69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b not found: ID does not exist" containerID="69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.295871 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b"} err="failed to get container status \"69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b\": rpc error: code = NotFound desc = could not find container \"69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b\": container with ID starting with 69f610c2b60b861db4e2685fd1225787d29dfd8a356102d89bdb041d9ae5997b not found: ID does not exist" Mar 14 09:24:55 crc kubenswrapper[4843]: I0314 09:24:55.350177 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" path="/var/lib/kubelet/pods/61b8979b-1447-4cb8-b87e-348f713d002b/volumes" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.281902 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-r4hhn"] Mar 14 09:24:58 crc kubenswrapper[4843]: E0314 09:24:58.282716 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="registry-server" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.282733 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="registry-server" Mar 14 09:24:58 crc kubenswrapper[4843]: E0314 09:24:58.282750 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="extract-utilities" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.282758 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="extract-utilities" Mar 14 09:24:58 crc kubenswrapper[4843]: E0314 09:24:58.282774 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="extract-content" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.282781 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="extract-content" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.282906 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="61b8979b-1447-4cb8-b87e-348f713d002b" containerName="registry-server" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.283358 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.286152 4843 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qzpgv" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.286174 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.287752 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.352831 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3520f5cf-b66e-4922-b114-2e7748a9103e-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-r4hhn\" (UID: \"3520f5cf-b66e-4922-b114-2e7748a9103e\") " pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.352894 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc7rv\" (UniqueName: \"kubernetes.io/projected/3520f5cf-b66e-4922-b114-2e7748a9103e-kube-api-access-kc7rv\") pod \"cert-manager-cainjector-5545bd876-r4hhn\" (UID: \"3520f5cf-b66e-4922-b114-2e7748a9103e\") " pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.453618 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3520f5cf-b66e-4922-b114-2e7748a9103e-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-r4hhn\" (UID: \"3520f5cf-b66e-4922-b114-2e7748a9103e\") " pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.453678 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc7rv\" (UniqueName: \"kubernetes.io/projected/3520f5cf-b66e-4922-b114-2e7748a9103e-kube-api-access-kc7rv\") pod \"cert-manager-cainjector-5545bd876-r4hhn\" (UID: \"3520f5cf-b66e-4922-b114-2e7748a9103e\") " pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.577623 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-r4hhn"] Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.591721 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3520f5cf-b66e-4922-b114-2e7748a9103e-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-r4hhn\" (UID: \"3520f5cf-b66e-4922-b114-2e7748a9103e\") " pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.603925 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc7rv\" (UniqueName: \"kubernetes.io/projected/3520f5cf-b66e-4922-b114-2e7748a9103e-kube-api-access-kc7rv\") pod \"cert-manager-cainjector-5545bd876-r4hhn\" (UID: \"3520f5cf-b66e-4922-b114-2e7748a9103e\") " pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.604578 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.645118 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k6ngx"] Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.665998 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.684917 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6ngx"] Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.778192 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knk2l\" (UniqueName: \"kubernetes.io/projected/21eabe83-90b8-4370-9366-3a5cbe4221aa-kube-api-access-knk2l\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.778294 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-utilities\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.778373 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-catalog-content\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.879543 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-catalog-content\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.879755 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knk2l\" (UniqueName: \"kubernetes.io/projected/21eabe83-90b8-4370-9366-3a5cbe4221aa-kube-api-access-knk2l\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.879786 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-utilities\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.880164 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-catalog-content\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.880177 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-utilities\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:58 crc kubenswrapper[4843]: I0314 09:24:58.923203 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knk2l\" (UniqueName: \"kubernetes.io/projected/21eabe83-90b8-4370-9366-3a5cbe4221aa-kube-api-access-knk2l\") pod \"certified-operators-k6ngx\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.089461 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.109590 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-r4hhn"] Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.251019 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" event={"ID":"3520f5cf-b66e-4922-b114-2e7748a9103e","Type":"ContainerStarted","Data":"13dca82107404fa9246666cc26b36fde9bab570488036b7a0ed4567c27b161f0"} Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.475106 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-2rbp4"] Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.476150 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.486090 4843 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gnszm" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.490333 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-2rbp4"] Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.534404 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6ngx"] Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.590852 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b200cf6-c7bf-4503-9694-fb0f63d901cc-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-2rbp4\" (UID: \"9b200cf6-c7bf-4503-9694-fb0f63d901cc\") " pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.590938 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2r5c\" (UniqueName: \"kubernetes.io/projected/9b200cf6-c7bf-4503-9694-fb0f63d901cc-kube-api-access-m2r5c\") pod \"cert-manager-webhook-6888856db4-2rbp4\" (UID: \"9b200cf6-c7bf-4503-9694-fb0f63d901cc\") " pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.691980 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2r5c\" (UniqueName: \"kubernetes.io/projected/9b200cf6-c7bf-4503-9694-fb0f63d901cc-kube-api-access-m2r5c\") pod \"cert-manager-webhook-6888856db4-2rbp4\" (UID: \"9b200cf6-c7bf-4503-9694-fb0f63d901cc\") " pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.692075 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b200cf6-c7bf-4503-9694-fb0f63d901cc-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-2rbp4\" (UID: \"9b200cf6-c7bf-4503-9694-fb0f63d901cc\") " pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.716825 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b200cf6-c7bf-4503-9694-fb0f63d901cc-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-2rbp4\" (UID: \"9b200cf6-c7bf-4503-9694-fb0f63d901cc\") " pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.717006 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2r5c\" (UniqueName: \"kubernetes.io/projected/9b200cf6-c7bf-4503-9694-fb0f63d901cc-kube-api-access-m2r5c\") pod \"cert-manager-webhook-6888856db4-2rbp4\" (UID: \"9b200cf6-c7bf-4503-9694-fb0f63d901cc\") " pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:24:59 crc kubenswrapper[4843]: I0314 09:24:59.795578 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:25:00 crc kubenswrapper[4843]: I0314 09:25:00.087360 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-2rbp4"] Mar 14 09:25:00 crc kubenswrapper[4843]: I0314 09:25:00.258050 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" event={"ID":"9b200cf6-c7bf-4503-9694-fb0f63d901cc","Type":"ContainerStarted","Data":"896a92c745d9582ab946ea127169c8dec8e570c3a82f047807807ed182b85fe8"} Mar 14 09:25:00 crc kubenswrapper[4843]: I0314 09:25:00.259586 4843 generic.go:334] "Generic (PLEG): container finished" podID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerID="d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e" exitCode=0 Mar 14 09:25:00 crc kubenswrapper[4843]: I0314 09:25:00.259619 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ngx" event={"ID":"21eabe83-90b8-4370-9366-3a5cbe4221aa","Type":"ContainerDied","Data":"d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e"} Mar 14 09:25:00 crc kubenswrapper[4843]: I0314 09:25:00.259645 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ngx" event={"ID":"21eabe83-90b8-4370-9366-3a5cbe4221aa","Type":"ContainerStarted","Data":"c804d0386730d494df0caab11c321b6c40d069a2bf65a60fb83492c957908e16"} Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.392528 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nrsqp"] Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.398006 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.430090 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrsqp"] Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.516366 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hvgh\" (UniqueName: \"kubernetes.io/projected/0b08028d-db0e-475d-ad59-01718c6c195a-kube-api-access-9hvgh\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.516439 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-utilities\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.516464 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-catalog-content\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.618201 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hvgh\" (UniqueName: \"kubernetes.io/projected/0b08028d-db0e-475d-ad59-01718c6c195a-kube-api-access-9hvgh\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.618360 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-utilities\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.618392 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-catalog-content\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.618977 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-utilities\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.619062 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-catalog-content\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.646521 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hvgh\" (UniqueName: \"kubernetes.io/projected/0b08028d-db0e-475d-ad59-01718c6c195a-kube-api-access-9hvgh\") pod \"redhat-marketplace-nrsqp\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:01 crc kubenswrapper[4843]: I0314 09:25:01.717854 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:02 crc kubenswrapper[4843]: I0314 09:25:02.202152 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrsqp"] Mar 14 09:25:02 crc kubenswrapper[4843]: W0314 09:25:02.211027 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b08028d_db0e_475d_ad59_01718c6c195a.slice/crio-56a530008549e75f6817cf241b09b980f1be0e75fccc6c6de2b9e7c28e39eece WatchSource:0}: Error finding container 56a530008549e75f6817cf241b09b980f1be0e75fccc6c6de2b9e7c28e39eece: Status 404 returned error can't find the container with id 56a530008549e75f6817cf241b09b980f1be0e75fccc6c6de2b9e7c28e39eece Mar 14 09:25:02 crc kubenswrapper[4843]: I0314 09:25:02.318010 4843 generic.go:334] "Generic (PLEG): container finished" podID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerID="aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21" exitCode=0 Mar 14 09:25:02 crc kubenswrapper[4843]: I0314 09:25:02.318098 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ngx" event={"ID":"21eabe83-90b8-4370-9366-3a5cbe4221aa","Type":"ContainerDied","Data":"aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21"} Mar 14 09:25:02 crc kubenswrapper[4843]: I0314 09:25:02.322188 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrsqp" event={"ID":"0b08028d-db0e-475d-ad59-01718c6c195a","Type":"ContainerStarted","Data":"56a530008549e75f6817cf241b09b980f1be0e75fccc6c6de2b9e7c28e39eece"} Mar 14 09:25:03 crc kubenswrapper[4843]: I0314 09:25:03.335607 4843 generic.go:334] "Generic (PLEG): container finished" podID="0b08028d-db0e-475d-ad59-01718c6c195a" containerID="01ade46128fbe5f7d33696d9164ba8860b3cd6e53cdbf6e0a1d61344c8355218" exitCode=0 Mar 14 09:25:03 crc kubenswrapper[4843]: I0314 09:25:03.335715 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrsqp" event={"ID":"0b08028d-db0e-475d-ad59-01718c6c195a","Type":"ContainerDied","Data":"01ade46128fbe5f7d33696d9164ba8860b3cd6e53cdbf6e0a1d61344c8355218"} Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.367931 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" event={"ID":"3520f5cf-b66e-4922-b114-2e7748a9103e","Type":"ContainerStarted","Data":"bafc082c6fe4dd08078151f5280a91b23c26587ed8287cb3da7a09934e2a3cbb"} Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.369853 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" event={"ID":"9b200cf6-c7bf-4503-9694-fb0f63d901cc","Type":"ContainerStarted","Data":"116cafe154f78244c0c3dbce76d4047fee4ae72175614dffb3b92cafb5cab38f"} Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.369896 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.373215 4843 generic.go:334] "Generic (PLEG): container finished" podID="0b08028d-db0e-475d-ad59-01718c6c195a" containerID="ffe435315e87a593c3b800afd88bd5f4e9db1e866a6d22fc5f48628ccb4eb50e" exitCode=0 Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.373287 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrsqp" event={"ID":"0b08028d-db0e-475d-ad59-01718c6c195a","Type":"ContainerDied","Data":"ffe435315e87a593c3b800afd88bd5f4e9db1e866a6d22fc5f48628ccb4eb50e"} Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.381174 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ngx" event={"ID":"21eabe83-90b8-4370-9366-3a5cbe4221aa","Type":"ContainerStarted","Data":"4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873"} Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.392782 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-r4hhn" podStartSLOduration=2.151756409 podStartE2EDuration="7.39276002s" podCreationTimestamp="2026-03-14 09:24:58 +0000 UTC" firstStartedPulling="2026-03-14 09:24:59.127602991 +0000 UTC m=+866.440214119" lastFinishedPulling="2026-03-14 09:25:04.368606602 +0000 UTC m=+871.681217730" observedRunningTime="2026-03-14 09:25:05.387126417 +0000 UTC m=+872.699737555" watchObservedRunningTime="2026-03-14 09:25:05.39276002 +0000 UTC m=+872.705371148" Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.489708 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" podStartSLOduration=2.242750062 podStartE2EDuration="6.489689528s" podCreationTimestamp="2026-03-14 09:24:59 +0000 UTC" firstStartedPulling="2026-03-14 09:25:00.102692986 +0000 UTC m=+867.415304114" lastFinishedPulling="2026-03-14 09:25:04.349632452 +0000 UTC m=+871.662243580" observedRunningTime="2026-03-14 09:25:05.447361404 +0000 UTC m=+872.759972532" watchObservedRunningTime="2026-03-14 09:25:05.489689528 +0000 UTC m=+872.802300656" Mar 14 09:25:05 crc kubenswrapper[4843]: I0314 09:25:05.491863 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k6ngx" podStartSLOduration=3.507962496 podStartE2EDuration="7.491853698s" podCreationTimestamp="2026-03-14 09:24:58 +0000 UTC" firstStartedPulling="2026-03-14 09:25:00.261351997 +0000 UTC m=+867.573963125" lastFinishedPulling="2026-03-14 09:25:04.245243199 +0000 UTC m=+871.557854327" observedRunningTime="2026-03-14 09:25:05.485358805 +0000 UTC m=+872.797969953" watchObservedRunningTime="2026-03-14 09:25:05.491853698 +0000 UTC m=+872.804464826" Mar 14 09:25:06 crc kubenswrapper[4843]: I0314 09:25:06.389502 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrsqp" event={"ID":"0b08028d-db0e-475d-ad59-01718c6c195a","Type":"ContainerStarted","Data":"6d91ebcd728ee1de1b8294832687fda0372de7960cc7018b15b2443668cda1b1"} Mar 14 09:25:06 crc kubenswrapper[4843]: I0314 09:25:06.413174 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nrsqp" podStartSLOduration=3.828549561 podStartE2EDuration="5.413151s" podCreationTimestamp="2026-03-14 09:25:01 +0000 UTC" firstStartedPulling="2026-03-14 09:25:04.211600722 +0000 UTC m=+871.524211850" lastFinishedPulling="2026-03-14 09:25:05.796202161 +0000 UTC m=+873.108813289" observedRunningTime="2026-03-14 09:25:06.409404372 +0000 UTC m=+873.722015520" watchObservedRunningTime="2026-03-14 09:25:06.413151 +0000 UTC m=+873.725762128" Mar 14 09:25:09 crc kubenswrapper[4843]: I0314 09:25:09.089743 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:25:09 crc kubenswrapper[4843]: I0314 09:25:09.090325 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:25:09 crc kubenswrapper[4843]: I0314 09:25:09.146383 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:25:09 crc kubenswrapper[4843]: I0314 09:25:09.452961 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:25:09 crc kubenswrapper[4843]: I0314 09:25:09.798545 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-2rbp4" Mar 14 09:25:09 crc kubenswrapper[4843]: I0314 09:25:09.983709 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6ngx"] Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.419982 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k6ngx" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="registry-server" containerID="cri-o://4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873" gracePeriod=2 Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.719024 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.719395 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.772803 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.812909 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.958325 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-utilities\") pod \"21eabe83-90b8-4370-9366-3a5cbe4221aa\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.958415 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knk2l\" (UniqueName: \"kubernetes.io/projected/21eabe83-90b8-4370-9366-3a5cbe4221aa-kube-api-access-knk2l\") pod \"21eabe83-90b8-4370-9366-3a5cbe4221aa\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.958469 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-catalog-content\") pod \"21eabe83-90b8-4370-9366-3a5cbe4221aa\" (UID: \"21eabe83-90b8-4370-9366-3a5cbe4221aa\") " Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.959424 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-utilities" (OuterVolumeSpecName: "utilities") pod "21eabe83-90b8-4370-9366-3a5cbe4221aa" (UID: "21eabe83-90b8-4370-9366-3a5cbe4221aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:25:11 crc kubenswrapper[4843]: I0314 09:25:11.966692 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21eabe83-90b8-4370-9366-3a5cbe4221aa-kube-api-access-knk2l" (OuterVolumeSpecName: "kube-api-access-knk2l") pod "21eabe83-90b8-4370-9366-3a5cbe4221aa" (UID: "21eabe83-90b8-4370-9366-3a5cbe4221aa"). InnerVolumeSpecName "kube-api-access-knk2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.059839 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knk2l\" (UniqueName: \"kubernetes.io/projected/21eabe83-90b8-4370-9366-3a5cbe4221aa-kube-api-access-knk2l\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.059872 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.428958 4843 generic.go:334] "Generic (PLEG): container finished" podID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerID="4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873" exitCode=0 Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.429007 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ngx" event={"ID":"21eabe83-90b8-4370-9366-3a5cbe4221aa","Type":"ContainerDied","Data":"4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873"} Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.429038 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ngx" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.429063 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ngx" event={"ID":"21eabe83-90b8-4370-9366-3a5cbe4221aa","Type":"ContainerDied","Data":"c804d0386730d494df0caab11c321b6c40d069a2bf65a60fb83492c957908e16"} Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.429086 4843 scope.go:117] "RemoveContainer" containerID="4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.456378 4843 scope.go:117] "RemoveContainer" containerID="aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.475266 4843 scope.go:117] "RemoveContainer" containerID="d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.487667 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.508607 4843 scope.go:117] "RemoveContainer" containerID="4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873" Mar 14 09:25:12 crc kubenswrapper[4843]: E0314 09:25:12.509354 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873\": container with ID starting with 4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873 not found: ID does not exist" containerID="4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.509401 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873"} err="failed to get container status \"4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873\": rpc error: code = NotFound desc = could not find container \"4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873\": container with ID starting with 4192d340585b3ad5a441421826e19bd9375084f12ad940f439d5cc0b06beb873 not found: ID does not exist" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.509428 4843 scope.go:117] "RemoveContainer" containerID="aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21" Mar 14 09:25:12 crc kubenswrapper[4843]: E0314 09:25:12.509902 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21\": container with ID starting with aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21 not found: ID does not exist" containerID="aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.509952 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21"} err="failed to get container status \"aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21\": rpc error: code = NotFound desc = could not find container \"aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21\": container with ID starting with aad59ee0c98b4a0ad96f0136c4b3da8be0b644549e1db8fff9e3234877b65b21 not found: ID does not exist" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.509979 4843 scope.go:117] "RemoveContainer" containerID="d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e" Mar 14 09:25:12 crc kubenswrapper[4843]: E0314 09:25:12.510691 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e\": container with ID starting with d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e not found: ID does not exist" containerID="d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e" Mar 14 09:25:12 crc kubenswrapper[4843]: I0314 09:25:12.510730 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e"} err="failed to get container status \"d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e\": rpc error: code = NotFound desc = could not find container \"d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e\": container with ID starting with d70e5bf41479e09ea784e36093fc666d9fd9659a9d78d156805898183153687e not found: ID does not exist" Mar 14 09:25:13 crc kubenswrapper[4843]: I0314 09:25:13.034959 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21eabe83-90b8-4370-9366-3a5cbe4221aa" (UID: "21eabe83-90b8-4370-9366-3a5cbe4221aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:25:13 crc kubenswrapper[4843]: I0314 09:25:13.073535 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21eabe83-90b8-4370-9366-3a5cbe4221aa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:13 crc kubenswrapper[4843]: I0314 09:25:13.375379 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6ngx"] Mar 14 09:25:13 crc kubenswrapper[4843]: I0314 09:25:13.380808 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k6ngx"] Mar 14 09:25:14 crc kubenswrapper[4843]: I0314 09:25:14.784390 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrsqp"] Mar 14 09:25:14 crc kubenswrapper[4843]: I0314 09:25:14.784608 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nrsqp" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="registry-server" containerID="cri-o://6d91ebcd728ee1de1b8294832687fda0372de7960cc7018b15b2443668cda1b1" gracePeriod=2 Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.308879 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-d6lvd"] Mar 14 09:25:15 crc kubenswrapper[4843]: E0314 09:25:15.309481 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="registry-server" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.309494 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="registry-server" Mar 14 09:25:15 crc kubenswrapper[4843]: E0314 09:25:15.309503 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="extract-utilities" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.309508 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="extract-utilities" Mar 14 09:25:15 crc kubenswrapper[4843]: E0314 09:25:15.309519 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="extract-content" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.309525 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="extract-content" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.309629 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" containerName="registry-server" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.310021 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.312854 4843 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vddhc" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.323066 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-d6lvd"] Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.347479 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21eabe83-90b8-4370-9366-3a5cbe4221aa" path="/var/lib/kubelet/pods/21eabe83-90b8-4370-9366-3a5cbe4221aa/volumes" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.402411 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zm5n\" (UniqueName: \"kubernetes.io/projected/94bce3e7-2401-4f5c-919f-6cadeb6472f4-kube-api-access-5zm5n\") pod \"cert-manager-545d4d4674-d6lvd\" (UID: \"94bce3e7-2401-4f5c-919f-6cadeb6472f4\") " pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.402692 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94bce3e7-2401-4f5c-919f-6cadeb6472f4-bound-sa-token\") pod \"cert-manager-545d4d4674-d6lvd\" (UID: \"94bce3e7-2401-4f5c-919f-6cadeb6472f4\") " pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.454497 4843 generic.go:334] "Generic (PLEG): container finished" podID="0b08028d-db0e-475d-ad59-01718c6c195a" containerID="6d91ebcd728ee1de1b8294832687fda0372de7960cc7018b15b2443668cda1b1" exitCode=0 Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.454512 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrsqp" event={"ID":"0b08028d-db0e-475d-ad59-01718c6c195a","Type":"ContainerDied","Data":"6d91ebcd728ee1de1b8294832687fda0372de7960cc7018b15b2443668cda1b1"} Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.504362 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zm5n\" (UniqueName: \"kubernetes.io/projected/94bce3e7-2401-4f5c-919f-6cadeb6472f4-kube-api-access-5zm5n\") pod \"cert-manager-545d4d4674-d6lvd\" (UID: \"94bce3e7-2401-4f5c-919f-6cadeb6472f4\") " pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.504436 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94bce3e7-2401-4f5c-919f-6cadeb6472f4-bound-sa-token\") pod \"cert-manager-545d4d4674-d6lvd\" (UID: \"94bce3e7-2401-4f5c-919f-6cadeb6472f4\") " pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.526011 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94bce3e7-2401-4f5c-919f-6cadeb6472f4-bound-sa-token\") pod \"cert-manager-545d4d4674-d6lvd\" (UID: \"94bce3e7-2401-4f5c-919f-6cadeb6472f4\") " pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.526127 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zm5n\" (UniqueName: \"kubernetes.io/projected/94bce3e7-2401-4f5c-919f-6cadeb6472f4-kube-api-access-5zm5n\") pod \"cert-manager-545d4d4674-d6lvd\" (UID: \"94bce3e7-2401-4f5c-919f-6cadeb6472f4\") " pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:15 crc kubenswrapper[4843]: I0314 09:25:15.626195 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-d6lvd" Mar 14 09:25:16 crc kubenswrapper[4843]: I0314 09:25:16.030820 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-d6lvd"] Mar 14 09:25:16 crc kubenswrapper[4843]: W0314 09:25:16.034428 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94bce3e7_2401_4f5c_919f_6cadeb6472f4.slice/crio-af493aa339289f0ee08352b445d68485c4ad504738f6164edc61d93668ca4778 WatchSource:0}: Error finding container af493aa339289f0ee08352b445d68485c4ad504738f6164edc61d93668ca4778: Status 404 returned error can't find the container with id af493aa339289f0ee08352b445d68485c4ad504738f6164edc61d93668ca4778 Mar 14 09:25:16 crc kubenswrapper[4843]: I0314 09:25:16.468815 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-d6lvd" event={"ID":"94bce3e7-2401-4f5c-919f-6cadeb6472f4","Type":"ContainerStarted","Data":"af493aa339289f0ee08352b445d68485c4ad504738f6164edc61d93668ca4778"} Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.479030 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-d6lvd" event={"ID":"94bce3e7-2401-4f5c-919f-6cadeb6472f4","Type":"ContainerStarted","Data":"7b13958ece5f45233add419e3b9a34f43c5c99e64161bb0c2e74fa85e572d3a3"} Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.500228 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-d6lvd" podStartSLOduration=2.50020241 podStartE2EDuration="2.50020241s" podCreationTimestamp="2026-03-14 09:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:25:17.498551311 +0000 UTC m=+884.811162469" watchObservedRunningTime="2026-03-14 09:25:17.50020241 +0000 UTC m=+884.812813558" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.610094 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.737065 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-catalog-content\") pod \"0b08028d-db0e-475d-ad59-01718c6c195a\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.737512 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-utilities\") pod \"0b08028d-db0e-475d-ad59-01718c6c195a\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.737593 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hvgh\" (UniqueName: \"kubernetes.io/projected/0b08028d-db0e-475d-ad59-01718c6c195a-kube-api-access-9hvgh\") pod \"0b08028d-db0e-475d-ad59-01718c6c195a\" (UID: \"0b08028d-db0e-475d-ad59-01718c6c195a\") " Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.738497 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-utilities" (OuterVolumeSpecName: "utilities") pod "0b08028d-db0e-475d-ad59-01718c6c195a" (UID: "0b08028d-db0e-475d-ad59-01718c6c195a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.743993 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b08028d-db0e-475d-ad59-01718c6c195a-kube-api-access-9hvgh" (OuterVolumeSpecName: "kube-api-access-9hvgh") pod "0b08028d-db0e-475d-ad59-01718c6c195a" (UID: "0b08028d-db0e-475d-ad59-01718c6c195a"). InnerVolumeSpecName "kube-api-access-9hvgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.769689 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b08028d-db0e-475d-ad59-01718c6c195a" (UID: "0b08028d-db0e-475d-ad59-01718c6c195a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.839089 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.839149 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hvgh\" (UniqueName: \"kubernetes.io/projected/0b08028d-db0e-475d-ad59-01718c6c195a-kube-api-access-9hvgh\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:17 crc kubenswrapper[4843]: I0314 09:25:17.839163 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08028d-db0e-475d-ad59-01718c6c195a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.490787 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrsqp" event={"ID":"0b08028d-db0e-475d-ad59-01718c6c195a","Type":"ContainerDied","Data":"56a530008549e75f6817cf241b09b980f1be0e75fccc6c6de2b9e7c28e39eece"} Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.490816 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrsqp" Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.491480 4843 scope.go:117] "RemoveContainer" containerID="6d91ebcd728ee1de1b8294832687fda0372de7960cc7018b15b2443668cda1b1" Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.522916 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrsqp"] Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.523558 4843 scope.go:117] "RemoveContainer" containerID="ffe435315e87a593c3b800afd88bd5f4e9db1e866a6d22fc5f48628ccb4eb50e" Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.529148 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrsqp"] Mar 14 09:25:18 crc kubenswrapper[4843]: I0314 09:25:18.541412 4843 scope.go:117] "RemoveContainer" containerID="01ade46128fbe5f7d33696d9164ba8860b3cd6e53cdbf6e0a1d61344c8355218" Mar 14 09:25:19 crc kubenswrapper[4843]: I0314 09:25:19.347030 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" path="/var/lib/kubelet/pods/0b08028d-db0e-475d-ad59-01718c6c195a/volumes" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.204197 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-72z8w"] Mar 14 09:25:23 crc kubenswrapper[4843]: E0314 09:25:23.205426 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="extract-content" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.205457 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="extract-content" Mar 14 09:25:23 crc kubenswrapper[4843]: E0314 09:25:23.205480 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="extract-utilities" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.205496 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="extract-utilities" Mar 14 09:25:23 crc kubenswrapper[4843]: E0314 09:25:23.205529 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="registry-server" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.205546 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="registry-server" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.205835 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b08028d-db0e-475d-ad59-01718c6c195a" containerName="registry-server" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.206977 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.213318 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7w8r4" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.213371 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.213805 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.222542 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-72z8w"] Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.314484 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4gjr\" (UniqueName: \"kubernetes.io/projected/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2-kube-api-access-x4gjr\") pod \"openstack-operator-index-72z8w\" (UID: \"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2\") " pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.415468 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4gjr\" (UniqueName: \"kubernetes.io/projected/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2-kube-api-access-x4gjr\") pod \"openstack-operator-index-72z8w\" (UID: \"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2\") " pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.458093 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4gjr\" (UniqueName: \"kubernetes.io/projected/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2-kube-api-access-x4gjr\") pod \"openstack-operator-index-72z8w\" (UID: \"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2\") " pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.542939 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:23 crc kubenswrapper[4843]: I0314 09:25:23.973375 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-72z8w"] Mar 14 09:25:24 crc kubenswrapper[4843]: I0314 09:25:24.544346 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-72z8w" event={"ID":"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2","Type":"ContainerStarted","Data":"20d100e9c55240716da667cb02763b4f820e03967d9844aa85c9ff6b70ce7ed2"} Mar 14 09:25:26 crc kubenswrapper[4843]: I0314 09:25:26.382160 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-72z8w"] Mar 14 09:25:26 crc kubenswrapper[4843]: I0314 09:25:26.983976 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6s6vv"] Mar 14 09:25:26 crc kubenswrapper[4843]: I0314 09:25:26.985251 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:26 crc kubenswrapper[4843]: I0314 09:25:26.993074 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6s6vv"] Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.063488 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfxms\" (UniqueName: \"kubernetes.io/projected/8f507af6-440f-42c3-839b-0466d546a5a0-kube-api-access-mfxms\") pod \"openstack-operator-index-6s6vv\" (UID: \"8f507af6-440f-42c3-839b-0466d546a5a0\") " pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.164919 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfxms\" (UniqueName: \"kubernetes.io/projected/8f507af6-440f-42c3-839b-0466d546a5a0-kube-api-access-mfxms\") pod \"openstack-operator-index-6s6vv\" (UID: \"8f507af6-440f-42c3-839b-0466d546a5a0\") " pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.198094 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfxms\" (UniqueName: \"kubernetes.io/projected/8f507af6-440f-42c3-839b-0466d546a5a0-kube-api-access-mfxms\") pod \"openstack-operator-index-6s6vv\" (UID: \"8f507af6-440f-42c3-839b-0466d546a5a0\") " pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.304387 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.567972 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-72z8w" event={"ID":"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2","Type":"ContainerStarted","Data":"9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584"} Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.568097 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-72z8w" podUID="2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" containerName="registry-server" containerID="cri-o://9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584" gracePeriod=2 Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.588242 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-72z8w" podStartSLOduration=1.789207252 podStartE2EDuration="4.588219057s" podCreationTimestamp="2026-03-14 09:25:23 +0000 UTC" firstStartedPulling="2026-03-14 09:25:23.986888 +0000 UTC m=+891.299499138" lastFinishedPulling="2026-03-14 09:25:26.785899805 +0000 UTC m=+894.098510943" observedRunningTime="2026-03-14 09:25:27.581124859 +0000 UTC m=+894.893735987" watchObservedRunningTime="2026-03-14 09:25:27.588219057 +0000 UTC m=+894.900830175" Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.743050 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6s6vv"] Mar 14 09:25:27 crc kubenswrapper[4843]: W0314 09:25:27.755299 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f507af6_440f_42c3_839b_0466d546a5a0.slice/crio-9f3dc36d8e6e552114540ceb3e30e9e9040b944b2e5cf4976a9b7ae36dfead9e WatchSource:0}: Error finding container 9f3dc36d8e6e552114540ceb3e30e9e9040b944b2e5cf4976a9b7ae36dfead9e: Status 404 returned error can't find the container with id 9f3dc36d8e6e552114540ceb3e30e9e9040b944b2e5cf4976a9b7ae36dfead9e Mar 14 09:25:27 crc kubenswrapper[4843]: I0314 09:25:27.995911 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.077679 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4gjr\" (UniqueName: \"kubernetes.io/projected/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2-kube-api-access-x4gjr\") pod \"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2\" (UID: \"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2\") " Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.089507 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2-kube-api-access-x4gjr" (OuterVolumeSpecName: "kube-api-access-x4gjr") pod "2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" (UID: "2b5bacf0-a6b4-4f30-b789-12b1cb651cf2"). InnerVolumeSpecName "kube-api-access-x4gjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.179899 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4gjr\" (UniqueName: \"kubernetes.io/projected/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2-kube-api-access-x4gjr\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.583849 4843 generic.go:334] "Generic (PLEG): container finished" podID="2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" containerID="9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584" exitCode=0 Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.583894 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-72z8w" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.583917 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-72z8w" event={"ID":"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2","Type":"ContainerDied","Data":"9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584"} Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.583944 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-72z8w" event={"ID":"2b5bacf0-a6b4-4f30-b789-12b1cb651cf2","Type":"ContainerDied","Data":"20d100e9c55240716da667cb02763b4f820e03967d9844aa85c9ff6b70ce7ed2"} Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.583965 4843 scope.go:117] "RemoveContainer" containerID="9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.587916 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6s6vv" event={"ID":"8f507af6-440f-42c3-839b-0466d546a5a0","Type":"ContainerStarted","Data":"5200e857daa705ee236a2d195c217dab0ea26728d661b62852b5e218878e7000"} Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.587994 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6s6vv" event={"ID":"8f507af6-440f-42c3-839b-0466d546a5a0","Type":"ContainerStarted","Data":"9f3dc36d8e6e552114540ceb3e30e9e9040b944b2e5cf4976a9b7ae36dfead9e"} Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.612414 4843 scope.go:117] "RemoveContainer" containerID="9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584" Mar 14 09:25:28 crc kubenswrapper[4843]: E0314 09:25:28.612942 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584\": container with ID starting with 9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584 not found: ID does not exist" containerID="9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.612977 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584"} err="failed to get container status \"9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584\": rpc error: code = NotFound desc = could not find container \"9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584\": container with ID starting with 9c8cef470810a25f5fed038fd814281ee1123df7e88d67d1b2be3642ea7df584 not found: ID does not exist" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.613843 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6s6vv" podStartSLOduration=2.545533991 podStartE2EDuration="2.613823159s" podCreationTimestamp="2026-03-14 09:25:26 +0000 UTC" firstStartedPulling="2026-03-14 09:25:27.763129142 +0000 UTC m=+895.075740280" lastFinishedPulling="2026-03-14 09:25:27.83141833 +0000 UTC m=+895.144029448" observedRunningTime="2026-03-14 09:25:28.609095978 +0000 UTC m=+895.921707116" watchObservedRunningTime="2026-03-14 09:25:28.613823159 +0000 UTC m=+895.926434297" Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.630529 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-72z8w"] Mar 14 09:25:28 crc kubenswrapper[4843]: I0314 09:25:28.648753 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-72z8w"] Mar 14 09:25:29 crc kubenswrapper[4843]: I0314 09:25:29.349028 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" path="/var/lib/kubelet/pods/2b5bacf0-a6b4-4f30-b789-12b1cb651cf2/volumes" Mar 14 09:25:37 crc kubenswrapper[4843]: I0314 09:25:37.304688 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:37 crc kubenswrapper[4843]: I0314 09:25:37.305297 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:37 crc kubenswrapper[4843]: I0314 09:25:37.351054 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:37 crc kubenswrapper[4843]: I0314 09:25:37.709930 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-6s6vv" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.677878 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7"] Mar 14 09:25:43 crc kubenswrapper[4843]: E0314 09:25:43.679169 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" containerName="registry-server" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.679201 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" containerName="registry-server" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.679560 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b5bacf0-a6b4-4f30-b789-12b1cb651cf2" containerName="registry-server" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.681529 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.684576 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7m7z9" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.687514 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7"] Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.800528 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-util\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.800582 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-bundle\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.800702 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfmfl\" (UniqueName: \"kubernetes.io/projected/9892ab03-eb13-4560-a688-5e38fd015478-kube-api-access-gfmfl\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.902223 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-bundle\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.902408 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfmfl\" (UniqueName: \"kubernetes.io/projected/9892ab03-eb13-4560-a688-5e38fd015478-kube-api-access-gfmfl\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.902571 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-util\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.902634 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-bundle\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.903563 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-util\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:43 crc kubenswrapper[4843]: I0314 09:25:43.928411 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfmfl\" (UniqueName: \"kubernetes.io/projected/9892ab03-eb13-4560-a688-5e38fd015478-kube-api-access-gfmfl\") pod \"b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:44 crc kubenswrapper[4843]: I0314 09:25:44.009026 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:44 crc kubenswrapper[4843]: I0314 09:25:44.500333 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7"] Mar 14 09:25:44 crc kubenswrapper[4843]: W0314 09:25:44.512234 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9892ab03_eb13_4560_a688_5e38fd015478.slice/crio-cbfcf375115a2b5826ca2842dd806a7e59e85f01732d078d5c20be142d996c6e WatchSource:0}: Error finding container cbfcf375115a2b5826ca2842dd806a7e59e85f01732d078d5c20be142d996c6e: Status 404 returned error can't find the container with id cbfcf375115a2b5826ca2842dd806a7e59e85f01732d078d5c20be142d996c6e Mar 14 09:25:44 crc kubenswrapper[4843]: I0314 09:25:44.723421 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" event={"ID":"9892ab03-eb13-4560-a688-5e38fd015478","Type":"ContainerStarted","Data":"8975b1669e8c0b88d662257e3f1f8d6348788475ce6143fddad2b2b63db8aeae"} Mar 14 09:25:44 crc kubenswrapper[4843]: I0314 09:25:44.723482 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" event={"ID":"9892ab03-eb13-4560-a688-5e38fd015478","Type":"ContainerStarted","Data":"cbfcf375115a2b5826ca2842dd806a7e59e85f01732d078d5c20be142d996c6e"} Mar 14 09:25:45 crc kubenswrapper[4843]: I0314 09:25:45.733490 4843 generic.go:334] "Generic (PLEG): container finished" podID="9892ab03-eb13-4560-a688-5e38fd015478" containerID="8975b1669e8c0b88d662257e3f1f8d6348788475ce6143fddad2b2b63db8aeae" exitCode=0 Mar 14 09:25:45 crc kubenswrapper[4843]: I0314 09:25:45.733556 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" event={"ID":"9892ab03-eb13-4560-a688-5e38fd015478","Type":"ContainerDied","Data":"8975b1669e8c0b88d662257e3f1f8d6348788475ce6143fddad2b2b63db8aeae"} Mar 14 09:25:46 crc kubenswrapper[4843]: I0314 09:25:46.749764 4843 generic.go:334] "Generic (PLEG): container finished" podID="9892ab03-eb13-4560-a688-5e38fd015478" containerID="78319b1ae76ee5fcc1f21965e4d906d61a6c9d5c614994662cdd60d90b7c86bc" exitCode=0 Mar 14 09:25:46 crc kubenswrapper[4843]: I0314 09:25:46.750112 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" event={"ID":"9892ab03-eb13-4560-a688-5e38fd015478","Type":"ContainerDied","Data":"78319b1ae76ee5fcc1f21965e4d906d61a6c9d5c614994662cdd60d90b7c86bc"} Mar 14 09:25:47 crc kubenswrapper[4843]: I0314 09:25:47.759744 4843 generic.go:334] "Generic (PLEG): container finished" podID="9892ab03-eb13-4560-a688-5e38fd015478" containerID="afa85199162200104443ef8465544e1a1e3403e9ee57444677ebb7b87e7542a3" exitCode=0 Mar 14 09:25:47 crc kubenswrapper[4843]: I0314 09:25:47.759815 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" event={"ID":"9892ab03-eb13-4560-a688-5e38fd015478","Type":"ContainerDied","Data":"afa85199162200104443ef8465544e1a1e3403e9ee57444677ebb7b87e7542a3"} Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.048688 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.210508 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-bundle\") pod \"9892ab03-eb13-4560-a688-5e38fd015478\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.210599 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-util\") pod \"9892ab03-eb13-4560-a688-5e38fd015478\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.211911 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-bundle" (OuterVolumeSpecName: "bundle") pod "9892ab03-eb13-4560-a688-5e38fd015478" (UID: "9892ab03-eb13-4560-a688-5e38fd015478"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.213484 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfmfl\" (UniqueName: \"kubernetes.io/projected/9892ab03-eb13-4560-a688-5e38fd015478-kube-api-access-gfmfl\") pod \"9892ab03-eb13-4560-a688-5e38fd015478\" (UID: \"9892ab03-eb13-4560-a688-5e38fd015478\") " Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.214140 4843 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.221577 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9892ab03-eb13-4560-a688-5e38fd015478-kube-api-access-gfmfl" (OuterVolumeSpecName: "kube-api-access-gfmfl") pod "9892ab03-eb13-4560-a688-5e38fd015478" (UID: "9892ab03-eb13-4560-a688-5e38fd015478"). InnerVolumeSpecName "kube-api-access-gfmfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.229763 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-util" (OuterVolumeSpecName: "util") pod "9892ab03-eb13-4560-a688-5e38fd015478" (UID: "9892ab03-eb13-4560-a688-5e38fd015478"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.316041 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfmfl\" (UniqueName: \"kubernetes.io/projected/9892ab03-eb13-4560-a688-5e38fd015478-kube-api-access-gfmfl\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.316116 4843 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9892ab03-eb13-4560-a688-5e38fd015478-util\") on node \"crc\" DevicePath \"\"" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.783077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" event={"ID":"9892ab03-eb13-4560-a688-5e38fd015478","Type":"ContainerDied","Data":"cbfcf375115a2b5826ca2842dd806a7e59e85f01732d078d5c20be142d996c6e"} Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.783142 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbfcf375115a2b5826ca2842dd806a7e59e85f01732d078d5c20be142d996c6e" Mar 14 09:25:49 crc kubenswrapper[4843]: I0314 09:25:49.783163 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.416698 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp"] Mar 14 09:25:56 crc kubenswrapper[4843]: E0314 09:25:56.417491 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="extract" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.417504 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="extract" Mar 14 09:25:56 crc kubenswrapper[4843]: E0314 09:25:56.417516 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="pull" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.417521 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="pull" Mar 14 09:25:56 crc kubenswrapper[4843]: E0314 09:25:56.417543 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="util" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.417549 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="util" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.417665 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="9892ab03-eb13-4560-a688-5e38fd015478" containerName="extract" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.418091 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.420675 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-ssf76" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.442425 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp"] Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.526982 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2qfb\" (UniqueName: \"kubernetes.io/projected/af58a715-edeb-480c-98a1-f42077825287-kube-api-access-w2qfb\") pod \"openstack-operator-controller-init-6bdb46c895-cstpp\" (UID: \"af58a715-edeb-480c-98a1-f42077825287\") " pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.628622 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2qfb\" (UniqueName: \"kubernetes.io/projected/af58a715-edeb-480c-98a1-f42077825287-kube-api-access-w2qfb\") pod \"openstack-operator-controller-init-6bdb46c895-cstpp\" (UID: \"af58a715-edeb-480c-98a1-f42077825287\") " pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.653598 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2qfb\" (UniqueName: \"kubernetes.io/projected/af58a715-edeb-480c-98a1-f42077825287-kube-api-access-w2qfb\") pod \"openstack-operator-controller-init-6bdb46c895-cstpp\" (UID: \"af58a715-edeb-480c-98a1-f42077825287\") " pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.737640 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:25:56 crc kubenswrapper[4843]: I0314 09:25:56.953794 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp"] Mar 14 09:25:56 crc kubenswrapper[4843]: W0314 09:25:56.960029 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf58a715_edeb_480c_98a1_f42077825287.slice/crio-115af8b17df258c7634dc8eb91aa7b4cf2d3041b17bf3ccefb1863485418a90c WatchSource:0}: Error finding container 115af8b17df258c7634dc8eb91aa7b4cf2d3041b17bf3ccefb1863485418a90c: Status 404 returned error can't find the container with id 115af8b17df258c7634dc8eb91aa7b4cf2d3041b17bf3ccefb1863485418a90c Mar 14 09:25:57 crc kubenswrapper[4843]: I0314 09:25:57.866663 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" event={"ID":"af58a715-edeb-480c-98a1-f42077825287","Type":"ContainerStarted","Data":"115af8b17df258c7634dc8eb91aa7b4cf2d3041b17bf3ccefb1863485418a90c"} Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.134060 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558006-9fnwg"] Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.135413 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.139390 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.139728 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.140009 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.147858 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558006-9fnwg"] Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.190258 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cgpf\" (UniqueName: \"kubernetes.io/projected/f044e544-ec72-408a-9848-b98f70c0669a-kube-api-access-7cgpf\") pod \"auto-csr-approver-29558006-9fnwg\" (UID: \"f044e544-ec72-408a-9848-b98f70c0669a\") " pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.292079 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cgpf\" (UniqueName: \"kubernetes.io/projected/f044e544-ec72-408a-9848-b98f70c0669a-kube-api-access-7cgpf\") pod \"auto-csr-approver-29558006-9fnwg\" (UID: \"f044e544-ec72-408a-9848-b98f70c0669a\") " pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.313959 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cgpf\" (UniqueName: \"kubernetes.io/projected/f044e544-ec72-408a-9848-b98f70c0669a-kube-api-access-7cgpf\") pod \"auto-csr-approver-29558006-9fnwg\" (UID: \"f044e544-ec72-408a-9848-b98f70c0669a\") " pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:00 crc kubenswrapper[4843]: I0314 09:26:00.457354 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:01 crc kubenswrapper[4843]: I0314 09:26:01.566576 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558006-9fnwg"] Mar 14 09:26:01 crc kubenswrapper[4843]: I0314 09:26:01.892556 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" event={"ID":"f044e544-ec72-408a-9848-b98f70c0669a","Type":"ContainerStarted","Data":"8dcae2ea7d06c1c7e9688924c4925ac90f7abc9f7e5eb9e92afcbd386cb67a57"} Mar 14 09:26:01 crc kubenswrapper[4843]: I0314 09:26:01.893959 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" event={"ID":"af58a715-edeb-480c-98a1-f42077825287","Type":"ContainerStarted","Data":"da3c83f872d4062855f44c6aa2a2ecdda7254b497fb9af510c75abca102cc7fe"} Mar 14 09:26:01 crc kubenswrapper[4843]: I0314 09:26:01.894093 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:26:01 crc kubenswrapper[4843]: I0314 09:26:01.925743 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" podStartSLOduration=1.437424813 podStartE2EDuration="5.925725638s" podCreationTimestamp="2026-03-14 09:25:56 +0000 UTC" firstStartedPulling="2026-03-14 09:25:56.96166963 +0000 UTC m=+924.274280758" lastFinishedPulling="2026-03-14 09:26:01.449970455 +0000 UTC m=+928.762581583" observedRunningTime="2026-03-14 09:26:01.92452217 +0000 UTC m=+929.237133308" watchObservedRunningTime="2026-03-14 09:26:01.925725638 +0000 UTC m=+929.238336766" Mar 14 09:26:02 crc kubenswrapper[4843]: I0314 09:26:02.901821 4843 generic.go:334] "Generic (PLEG): container finished" podID="f044e544-ec72-408a-9848-b98f70c0669a" containerID="f1e0749317b316f6dbc387640e4947fc71f55e4904687cd9e916c67ef42dfa1b" exitCode=0 Mar 14 09:26:02 crc kubenswrapper[4843]: I0314 09:26:02.901896 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" event={"ID":"f044e544-ec72-408a-9848-b98f70c0669a","Type":"ContainerDied","Data":"f1e0749317b316f6dbc387640e4947fc71f55e4904687cd9e916c67ef42dfa1b"} Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.150682 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.248742 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cgpf\" (UniqueName: \"kubernetes.io/projected/f044e544-ec72-408a-9848-b98f70c0669a-kube-api-access-7cgpf\") pod \"f044e544-ec72-408a-9848-b98f70c0669a\" (UID: \"f044e544-ec72-408a-9848-b98f70c0669a\") " Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.254655 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f044e544-ec72-408a-9848-b98f70c0669a-kube-api-access-7cgpf" (OuterVolumeSpecName: "kube-api-access-7cgpf") pod "f044e544-ec72-408a-9848-b98f70c0669a" (UID: "f044e544-ec72-408a-9848-b98f70c0669a"). InnerVolumeSpecName "kube-api-access-7cgpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.350033 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cgpf\" (UniqueName: \"kubernetes.io/projected/f044e544-ec72-408a-9848-b98f70c0669a-kube-api-access-7cgpf\") on node \"crc\" DevicePath \"\"" Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.917842 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" event={"ID":"f044e544-ec72-408a-9848-b98f70c0669a","Type":"ContainerDied","Data":"8dcae2ea7d06c1c7e9688924c4925ac90f7abc9f7e5eb9e92afcbd386cb67a57"} Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.917892 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dcae2ea7d06c1c7e9688924c4925ac90f7abc9f7e5eb9e92afcbd386cb67a57" Mar 14 09:26:04 crc kubenswrapper[4843]: I0314 09:26:04.917961 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558006-9fnwg" Mar 14 09:26:05 crc kubenswrapper[4843]: E0314 09:26:05.059387 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf044e544_ec72_408a_9848_b98f70c0669a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf044e544_ec72_408a_9848_b98f70c0669a.slice/crio-8dcae2ea7d06c1c7e9688924c4925ac90f7abc9f7e5eb9e92afcbd386cb67a57\": RecentStats: unable to find data in memory cache]" Mar 14 09:26:05 crc kubenswrapper[4843]: I0314 09:26:05.221415 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558000-dj6lc"] Mar 14 09:26:05 crc kubenswrapper[4843]: I0314 09:26:05.230886 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558000-dj6lc"] Mar 14 09:26:05 crc kubenswrapper[4843]: I0314 09:26:05.349648 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac4bb14-edec-4967-a71f-20e67f264a0a" path="/var/lib/kubelet/pods/9ac4bb14-edec-4967-a71f-20e67f264a0a/volumes" Mar 14 09:26:06 crc kubenswrapper[4843]: I0314 09:26:06.741517 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:26:15 crc kubenswrapper[4843]: I0314 09:26:15.387809 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:26:15 crc kubenswrapper[4843]: I0314 09:26:15.388409 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.083901 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z"] Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.084722 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f044e544-ec72-408a-9848-b98f70c0669a" containerName="oc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.084736 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f044e544-ec72-408a-9848-b98f70c0669a" containerName="oc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.084851 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f044e544-ec72-408a-9848-b98f70c0669a" containerName="oc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.085300 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.087153 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-8wd8x" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.094776 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.095941 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.099758 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jhjj4" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.107204 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.114164 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.115473 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.132172 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-vtt9d" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.140228 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.149053 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.157371 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.158190 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.162828 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-9bw4t" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.174748 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.175381 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdjp6\" (UniqueName: \"kubernetes.io/projected/0fbf4f22-f279-4841-b829-6d35d2dd0bfc-kube-api-access-vdjp6\") pod \"cinder-operator-controller-manager-984cd4dcf-75kt6\" (UID: \"0fbf4f22-f279-4841-b829-6d35d2dd0bfc\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.175454 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxk4f\" (UniqueName: \"kubernetes.io/projected/0e6cdd9e-39ed-4359-9b74-54e23a0c70e2-kube-api-access-jxk4f\") pod \"designate-operator-controller-manager-66d56f6ff4-mxn5j\" (UID: \"0e6cdd9e-39ed-4359-9b74-54e23a0c70e2\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.175484 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tdsz\" (UniqueName: \"kubernetes.io/projected/09b1d07d-69b7-427b-9577-2a782637a2b1-kube-api-access-4tdsz\") pod \"barbican-operator-controller-manager-d47688694-ktr4z\" (UID: \"09b1d07d-69b7-427b-9577-2a782637a2b1\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.175554 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n247\" (UniqueName: \"kubernetes.io/projected/354f7562-fd32-491f-a3dd-b42bcf7d30cb-kube-api-access-6n247\") pod \"glance-operator-controller-manager-5964f64c48-g29z7\" (UID: \"354f7562-fd32-491f-a3dd-b42bcf7d30cb\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.180342 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-czghc"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.181178 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.183923 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-gxv8b" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.211367 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-czghc"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.241116 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.242009 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.251024 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-zbnd8" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.255042 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.255890 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.264911 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.266618 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4jsrl" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.266724 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.273613 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.274411 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277052 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-7dnws" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277813 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxk4f\" (UniqueName: \"kubernetes.io/projected/0e6cdd9e-39ed-4359-9b74-54e23a0c70e2-kube-api-access-jxk4f\") pod \"designate-operator-controller-manager-66d56f6ff4-mxn5j\" (UID: \"0e6cdd9e-39ed-4359-9b74-54e23a0c70e2\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277848 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbsv9\" (UniqueName: \"kubernetes.io/projected/3da8acd2-64eb-4f3f-b28e-80d5bb4e4743-kube-api-access-wbsv9\") pod \"horizon-operator-controller-manager-6d9d6b584d-bmb27\" (UID: \"3da8acd2-64eb-4f3f-b28e-80d5bb4e4743\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277870 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7fr8\" (UniqueName: \"kubernetes.io/projected/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-kube-api-access-v7fr8\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277887 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tdsz\" (UniqueName: \"kubernetes.io/projected/09b1d07d-69b7-427b-9577-2a782637a2b1-kube-api-access-4tdsz\") pod \"barbican-operator-controller-manager-d47688694-ktr4z\" (UID: \"09b1d07d-69b7-427b-9577-2a782637a2b1\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277917 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277951 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddkk2\" (UniqueName: \"kubernetes.io/projected/7d4e28ff-8c1b-4533-a733-97c44a2b7b0a-kube-api-access-ddkk2\") pod \"heat-operator-controller-manager-77b6666d85-czghc\" (UID: \"7d4e28ff-8c1b-4533-a733-97c44a2b7b0a\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277974 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n247\" (UniqueName: \"kubernetes.io/projected/354f7562-fd32-491f-a3dd-b42bcf7d30cb-kube-api-access-6n247\") pod \"glance-operator-controller-manager-5964f64c48-g29z7\" (UID: \"354f7562-fd32-491f-a3dd-b42bcf7d30cb\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.277997 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdjp6\" (UniqueName: \"kubernetes.io/projected/0fbf4f22-f279-4841-b829-6d35d2dd0bfc-kube-api-access-vdjp6\") pod \"cinder-operator-controller-manager-984cd4dcf-75kt6\" (UID: \"0fbf4f22-f279-4841-b829-6d35d2dd0bfc\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.278904 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.292090 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.292853 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.297556 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-9v2nj" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.302415 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.322968 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdjp6\" (UniqueName: \"kubernetes.io/projected/0fbf4f22-f279-4841-b829-6d35d2dd0bfc-kube-api-access-vdjp6\") pod \"cinder-operator-controller-manager-984cd4dcf-75kt6\" (UID: \"0fbf4f22-f279-4841-b829-6d35d2dd0bfc\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.333497 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.338077 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxk4f\" (UniqueName: \"kubernetes.io/projected/0e6cdd9e-39ed-4359-9b74-54e23a0c70e2-kube-api-access-jxk4f\") pod \"designate-operator-controller-manager-66d56f6ff4-mxn5j\" (UID: \"0e6cdd9e-39ed-4359-9b74-54e23a0c70e2\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.342869 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n247\" (UniqueName: \"kubernetes.io/projected/354f7562-fd32-491f-a3dd-b42bcf7d30cb-kube-api-access-6n247\") pod \"glance-operator-controller-manager-5964f64c48-g29z7\" (UID: \"354f7562-fd32-491f-a3dd-b42bcf7d30cb\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.346246 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tdsz\" (UniqueName: \"kubernetes.io/projected/09b1d07d-69b7-427b-9577-2a782637a2b1-kube-api-access-4tdsz\") pod \"barbican-operator-controller-manager-d47688694-ktr4z\" (UID: \"09b1d07d-69b7-427b-9577-2a782637a2b1\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.346361 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.347779 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.350467 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-7dmvn" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.368599 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.377558 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.378528 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379213 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8q6n\" (UniqueName: \"kubernetes.io/projected/47f2d7ba-aac2-44f4-9146-ca20192a08ab-kube-api-access-b8q6n\") pod \"ironic-operator-controller-manager-5bc894d9b-8wvhp\" (UID: \"47f2d7ba-aac2-44f4-9146-ca20192a08ab\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379310 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbsv9\" (UniqueName: \"kubernetes.io/projected/3da8acd2-64eb-4f3f-b28e-80d5bb4e4743-kube-api-access-wbsv9\") pod \"horizon-operator-controller-manager-6d9d6b584d-bmb27\" (UID: \"3da8acd2-64eb-4f3f-b28e-80d5bb4e4743\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379341 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7fr8\" (UniqueName: \"kubernetes.io/projected/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-kube-api-access-v7fr8\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379434 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379494 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddkk2\" (UniqueName: \"kubernetes.io/projected/7d4e28ff-8c1b-4533-a733-97c44a2b7b0a-kube-api-access-ddkk2\") pod \"heat-operator-controller-manager-77b6666d85-czghc\" (UID: \"7d4e28ff-8c1b-4533-a733-97c44a2b7b0a\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379545 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdwtv\" (UniqueName: \"kubernetes.io/projected/720b9936-b2c8-4b7b-ab8f-c8ecfda9052d-kube-api-access-jdwtv\") pod \"manila-operator-controller-manager-57b484b4df-fqfbh\" (UID: \"720b9936-b2c8-4b7b-ab8f-c8ecfda9052d\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.379574 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l9fj\" (UniqueName: \"kubernetes.io/projected/d4d69189-dfff-4f22-b978-4988c505df47-kube-api-access-6l9fj\") pod \"keystone-operator-controller-manager-684f77d66d-7v8h2\" (UID: \"d4d69189-dfff-4f22-b978-4988c505df47\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.381357 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-8m7mb" Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.381493 4843 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.381543 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert podName:46ec1c5f-0334-4d5d-b16a-8e01ffc76613 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:28.881522276 +0000 UTC m=+956.194133474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert") pod "infra-operator-controller-manager-54dc5b8f8d-ffxmw" (UID: "46ec1c5f-0334-4d5d-b16a-8e01ffc76613") : secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.403629 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.422027 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.443774 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.446530 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddkk2\" (UniqueName: \"kubernetes.io/projected/7d4e28ff-8c1b-4533-a733-97c44a2b7b0a-kube-api-access-ddkk2\") pod \"heat-operator-controller-manager-77b6666d85-czghc\" (UID: \"7d4e28ff-8c1b-4533-a733-97c44a2b7b0a\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.447589 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7fr8\" (UniqueName: \"kubernetes.io/projected/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-kube-api-access-v7fr8\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.454576 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.463011 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbsv9\" (UniqueName: \"kubernetes.io/projected/3da8acd2-64eb-4f3f-b28e-80d5bb4e4743-kube-api-access-wbsv9\") pod \"horizon-operator-controller-manager-6d9d6b584d-bmb27\" (UID: \"3da8acd2-64eb-4f3f-b28e-80d5bb4e4743\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.469671 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.472008 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.477146 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-h5mnf" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.478553 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.486023 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.487061 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpsbk\" (UniqueName: \"kubernetes.io/projected/ddd8e1af-7084-4685-96a9-ca4eaa9a4742-kube-api-access-xpsbk\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8\" (UID: \"ddd8e1af-7084-4685-96a9-ca4eaa9a4742\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.487132 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdwtv\" (UniqueName: \"kubernetes.io/projected/720b9936-b2c8-4b7b-ab8f-c8ecfda9052d-kube-api-access-jdwtv\") pod \"manila-operator-controller-manager-57b484b4df-fqfbh\" (UID: \"720b9936-b2c8-4b7b-ab8f-c8ecfda9052d\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.487179 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l9fj\" (UniqueName: \"kubernetes.io/projected/d4d69189-dfff-4f22-b978-4988c505df47-kube-api-access-6l9fj\") pod \"keystone-operator-controller-manager-684f77d66d-7v8h2\" (UID: \"d4d69189-dfff-4f22-b978-4988c505df47\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.487212 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhlrn\" (UniqueName: \"kubernetes.io/projected/28f29a72-e196-4962-87b3-cc49b8591cc0-kube-api-access-qhlrn\") pod \"neutron-operator-controller-manager-776c5696bf-9dlkn\" (UID: \"28f29a72-e196-4962-87b3-cc49b8591cc0\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.487239 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8q6n\" (UniqueName: \"kubernetes.io/projected/47f2d7ba-aac2-44f4-9146-ca20192a08ab-kube-api-access-b8q6n\") pod \"ironic-operator-controller-manager-5bc894d9b-8wvhp\" (UID: \"47f2d7ba-aac2-44f4-9146-ca20192a08ab\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.495815 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-g59td"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.496776 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.497552 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.501428 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-tn4sz" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.512655 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l9fj\" (UniqueName: \"kubernetes.io/projected/d4d69189-dfff-4f22-b978-4988c505df47-kube-api-access-6l9fj\") pod \"keystone-operator-controller-manager-684f77d66d-7v8h2\" (UID: \"d4d69189-dfff-4f22-b978-4988c505df47\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.516849 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.521770 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-g59td"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.521909 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.524829 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-d779p" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.527229 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.538913 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8q6n\" (UniqueName: \"kubernetes.io/projected/47f2d7ba-aac2-44f4-9146-ca20192a08ab-kube-api-access-b8q6n\") pod \"ironic-operator-controller-manager-5bc894d9b-8wvhp\" (UID: \"47f2d7ba-aac2-44f4-9146-ca20192a08ab\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.540779 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdwtv\" (UniqueName: \"kubernetes.io/projected/720b9936-b2c8-4b7b-ab8f-c8ecfda9052d-kube-api-access-jdwtv\") pod \"manila-operator-controller-manager-57b484b4df-fqfbh\" (UID: \"720b9936-b2c8-4b7b-ab8f-c8ecfda9052d\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.545092 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.546344 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.549176 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.553032 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-88shv" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.560003 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.567630 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.577436 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.578435 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.588435 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-w2d4h" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.588610 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.589886 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wswl\" (UniqueName: \"kubernetes.io/projected/f5a4b439-9571-4fdb-8796-3bd729646aa3-kube-api-access-9wswl\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.589977 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.590069 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ccmc\" (UniqueName: \"kubernetes.io/projected/bdefcc7b-402d-4c04-bc48-aa8b918d1b15-kube-api-access-9ccmc\") pod \"placement-operator-controller-manager-574d45c66c-tx8dt\" (UID: \"bdefcc7b-402d-4c04-bc48-aa8b918d1b15\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.590117 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdz9t\" (UniqueName: \"kubernetes.io/projected/abe14da3-e8a8-45f6-aebb-5df03c384a35-kube-api-access-jdz9t\") pod \"nova-operator-controller-manager-7f84474648-g59td\" (UID: \"abe14da3-e8a8-45f6-aebb-5df03c384a35\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.590216 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7dz8\" (UniqueName: \"kubernetes.io/projected/72b0bab9-36e7-49da-af7b-ff4fe9475545-kube-api-access-m7dz8\") pod \"octavia-operator-controller-manager-5f4f55cb5c-zrfcm\" (UID: \"72b0bab9-36e7-49da-af7b-ff4fe9475545\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.590256 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpsbk\" (UniqueName: \"kubernetes.io/projected/ddd8e1af-7084-4685-96a9-ca4eaa9a4742-kube-api-access-xpsbk\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8\" (UID: \"ddd8e1af-7084-4685-96a9-ca4eaa9a4742\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.590342 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhlrn\" (UniqueName: \"kubernetes.io/projected/28f29a72-e196-4962-87b3-cc49b8591cc0-kube-api-access-qhlrn\") pod \"neutron-operator-controller-manager-776c5696bf-9dlkn\" (UID: \"28f29a72-e196-4962-87b3-cc49b8591cc0\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.592838 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.597352 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.597638 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vf7t5" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.605776 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.613356 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.613918 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpsbk\" (UniqueName: \"kubernetes.io/projected/ddd8e1af-7084-4685-96a9-ca4eaa9a4742-kube-api-access-xpsbk\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8\" (UID: \"ddd8e1af-7084-4685-96a9-ca4eaa9a4742\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.614037 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhlrn\" (UniqueName: \"kubernetes.io/projected/28f29a72-e196-4962-87b3-cc49b8591cc0-kube-api-access-qhlrn\") pod \"neutron-operator-controller-manager-776c5696bf-9dlkn\" (UID: \"28f29a72-e196-4962-87b3-cc49b8591cc0\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.614648 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.620167 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.628475 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-q9h7j" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.628683 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.636492 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.639051 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.641727 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-lg8td" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.647820 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.672203 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.673181 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.677678 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.677787 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-xvvq9" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690800 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gppv9\" (UniqueName: \"kubernetes.io/projected/cf4074be-d697-47bf-a44c-57822afebb15-kube-api-access-gppv9\") pod \"swift-operator-controller-manager-7f9cc5dd44-kbcdm\" (UID: \"cf4074be-d697-47bf-a44c-57822afebb15\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690838 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690884 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75qh\" (UniqueName: \"kubernetes.io/projected/f1e0e04a-1194-4fa6-878f-318d360845d3-kube-api-access-s75qh\") pod \"telemetry-operator-controller-manager-6854b8b9d9-vgstg\" (UID: \"f1e0e04a-1194-4fa6-878f-318d360845d3\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690906 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ccmc\" (UniqueName: \"kubernetes.io/projected/bdefcc7b-402d-4c04-bc48-aa8b918d1b15-kube-api-access-9ccmc\") pod \"placement-operator-controller-manager-574d45c66c-tx8dt\" (UID: \"bdefcc7b-402d-4c04-bc48-aa8b918d1b15\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690936 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdz9t\" (UniqueName: \"kubernetes.io/projected/abe14da3-e8a8-45f6-aebb-5df03c384a35-kube-api-access-jdz9t\") pod \"nova-operator-controller-manager-7f84474648-g59td\" (UID: \"abe14da3-e8a8-45f6-aebb-5df03c384a35\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690954 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pmvh\" (UniqueName: \"kubernetes.io/projected/99c2bbc7-1885-4031-9478-a839c36bfbc5-kube-api-access-5pmvh\") pod \"test-operator-controller-manager-5c5cb9c4d7-9nkkk\" (UID: \"99c2bbc7-1885-4031-9478-a839c36bfbc5\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.690972 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7dz8\" (UniqueName: \"kubernetes.io/projected/72b0bab9-36e7-49da-af7b-ff4fe9475545-kube-api-access-m7dz8\") pod \"octavia-operator-controller-manager-5f4f55cb5c-zrfcm\" (UID: \"72b0bab9-36e7-49da-af7b-ff4fe9475545\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.691008 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4274k\" (UniqueName: \"kubernetes.io/projected/92eda9b1-e96f-4a71-93f8-a595c6f5117d-kube-api-access-4274k\") pod \"ovn-operator-controller-manager-bbc5b68f9-xctxr\" (UID: \"92eda9b1-e96f-4a71-93f8-a595c6f5117d\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.691037 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wswl\" (UniqueName: \"kubernetes.io/projected/f5a4b439-9571-4fdb-8796-3bd729646aa3-kube-api-access-9wswl\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.691509 4843 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.691544 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert podName:f5a4b439-9571-4fdb-8796-3bd729646aa3 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:29.191532812 +0000 UTC m=+956.504143940 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" (UID: "f5a4b439-9571-4fdb-8796-3bd729646aa3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.719250 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.733210 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wswl\" (UniqueName: \"kubernetes.io/projected/f5a4b439-9571-4fdb-8796-3bd729646aa3-kube-api-access-9wswl\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.737264 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdz9t\" (UniqueName: \"kubernetes.io/projected/abe14da3-e8a8-45f6-aebb-5df03c384a35-kube-api-access-jdz9t\") pod \"nova-operator-controller-manager-7f84474648-g59td\" (UID: \"abe14da3-e8a8-45f6-aebb-5df03c384a35\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.738201 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.739660 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.747415 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-vn982" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.749550 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ccmc\" (UniqueName: \"kubernetes.io/projected/bdefcc7b-402d-4c04-bc48-aa8b918d1b15-kube-api-access-9ccmc\") pod \"placement-operator-controller-manager-574d45c66c-tx8dt\" (UID: \"bdefcc7b-402d-4c04-bc48-aa8b918d1b15\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.796143 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gppv9\" (UniqueName: \"kubernetes.io/projected/cf4074be-d697-47bf-a44c-57822afebb15-kube-api-access-gppv9\") pod \"swift-operator-controller-manager-7f9cc5dd44-kbcdm\" (UID: \"cf4074be-d697-47bf-a44c-57822afebb15\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.796262 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75qh\" (UniqueName: \"kubernetes.io/projected/f1e0e04a-1194-4fa6-878f-318d360845d3-kube-api-access-s75qh\") pod \"telemetry-operator-controller-manager-6854b8b9d9-vgstg\" (UID: \"f1e0e04a-1194-4fa6-878f-318d360845d3\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.796353 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pmvh\" (UniqueName: \"kubernetes.io/projected/99c2bbc7-1885-4031-9478-a839c36bfbc5-kube-api-access-5pmvh\") pod \"test-operator-controller-manager-5c5cb9c4d7-9nkkk\" (UID: \"99c2bbc7-1885-4031-9478-a839c36bfbc5\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.796451 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4274k\" (UniqueName: \"kubernetes.io/projected/92eda9b1-e96f-4a71-93f8-a595c6f5117d-kube-api-access-4274k\") pod \"ovn-operator-controller-manager-bbc5b68f9-xctxr\" (UID: \"92eda9b1-e96f-4a71-93f8-a595c6f5117d\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.796496 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-854cr\" (UniqueName: \"kubernetes.io/projected/54b06835-eef3-4b4d-bf36-2714b438588f-kube-api-access-854cr\") pod \"watcher-operator-controller-manager-7b8d757b5d-2cr4c\" (UID: \"54b06835-eef3-4b4d-bf36-2714b438588f\") " pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.804400 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7dz8\" (UniqueName: \"kubernetes.io/projected/72b0bab9-36e7-49da-af7b-ff4fe9475545-kube-api-access-m7dz8\") pod \"octavia-operator-controller-manager-5f4f55cb5c-zrfcm\" (UID: \"72b0bab9-36e7-49da-af7b-ff4fe9475545\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.829613 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pmvh\" (UniqueName: \"kubernetes.io/projected/99c2bbc7-1885-4031-9478-a839c36bfbc5-kube-api-access-5pmvh\") pod \"test-operator-controller-manager-5c5cb9c4d7-9nkkk\" (UID: \"99c2bbc7-1885-4031-9478-a839c36bfbc5\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.829906 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.832843 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75qh\" (UniqueName: \"kubernetes.io/projected/f1e0e04a-1194-4fa6-878f-318d360845d3-kube-api-access-s75qh\") pod \"telemetry-operator-controller-manager-6854b8b9d9-vgstg\" (UID: \"f1e0e04a-1194-4fa6-878f-318d360845d3\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.833247 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gppv9\" (UniqueName: \"kubernetes.io/projected/cf4074be-d697-47bf-a44c-57822afebb15-kube-api-access-gppv9\") pod \"swift-operator-controller-manager-7f9cc5dd44-kbcdm\" (UID: \"cf4074be-d697-47bf-a44c-57822afebb15\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.838594 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.853227 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4274k\" (UniqueName: \"kubernetes.io/projected/92eda9b1-e96f-4a71-93f8-a595c6f5117d-kube-api-access-4274k\") pod \"ovn-operator-controller-manager-bbc5b68f9-xctxr\" (UID: \"92eda9b1-e96f-4a71-93f8-a595c6f5117d\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.867358 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c"] Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.900595 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.902036 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.906261 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-854cr\" (UniqueName: \"kubernetes.io/projected/54b06835-eef3-4b4d-bf36-2714b438588f-kube-api-access-854cr\") pod \"watcher-operator-controller-manager-7b8d757b5d-2cr4c\" (UID: \"54b06835-eef3-4b4d-bf36-2714b438588f\") " pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.906482 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.907901 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.911612 4843 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:28 crc kubenswrapper[4843]: E0314 09:26:28.911684 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert podName:46ec1c5f-0334-4d5d-b16a-8e01ffc76613 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:29.911657048 +0000 UTC m=+957.224268176 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert") pod "infra-operator-controller-manager-54dc5b8f8d-ffxmw" (UID: "46ec1c5f-0334-4d5d-b16a-8e01ffc76613") : secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.942382 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.959895 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.965109 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:26:28 crc kubenswrapper[4843]: I0314 09:26:28.995721 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-854cr\" (UniqueName: \"kubernetes.io/projected/54b06835-eef3-4b4d-bf36-2714b438588f-kube-api-access-854cr\") pod \"watcher-operator-controller-manager-7b8d757b5d-2cr4c\" (UID: \"54b06835-eef3-4b4d-bf36-2714b438588f\") " pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.002216 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.066013 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb"] Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.067997 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.070717 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.071851 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9t5km" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.072028 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.077044 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.079749 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb"] Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.117757 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78"] Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.118917 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.127229 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nvv54" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.131683 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.145187 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78"] Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.174256 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.174400 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.174473 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xjnx\" (UniqueName: \"kubernetes.io/projected/a0339dbb-a75e-465b-a30a-75c3091f506c-kube-api-access-6xjnx\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.275887 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.275955 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.275989 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xjnx\" (UniqueName: \"kubernetes.io/projected/a0339dbb-a75e-465b-a30a-75c3091f506c-kube-api-access-6xjnx\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.276010 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4nbl\" (UniqueName: \"kubernetes.io/projected/f75cfc62-a50d-4bad-9868-c58dc328cdbb-kube-api-access-c4nbl\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qjs78\" (UID: \"f75cfc62-a50d-4bad-9868-c58dc328cdbb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.276031 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.276194 4843 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.276242 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:29.776227137 +0000 UTC m=+957.088838265 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "metrics-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.276302 4843 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.276365 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert podName:f5a4b439-9571-4fdb-8796-3bd729646aa3 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:30.276346559 +0000 UTC m=+957.588957687 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" (UID: "f5a4b439-9571-4fdb-8796-3bd729646aa3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.276408 4843 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.276433 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:29.776424231 +0000 UTC m=+957.089035359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.311287 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xjnx\" (UniqueName: \"kubernetes.io/projected/a0339dbb-a75e-465b-a30a-75c3091f506c-kube-api-access-6xjnx\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.376791 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z"] Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.377409 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4nbl\" (UniqueName: \"kubernetes.io/projected/f75cfc62-a50d-4bad-9868-c58dc328cdbb-kube-api-access-c4nbl\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qjs78\" (UID: \"f75cfc62-a50d-4bad-9868-c58dc328cdbb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.406710 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4nbl\" (UniqueName: \"kubernetes.io/projected/f75cfc62-a50d-4bad-9868-c58dc328cdbb-kube-api-access-c4nbl\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qjs78\" (UID: \"f75cfc62-a50d-4bad-9868-c58dc328cdbb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.455625 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.499988 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j"] Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.786356 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.786442 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.786531 4843 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.786592 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:30.786575391 +0000 UTC m=+958.099186509 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.786609 4843 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.786685 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:30.786666693 +0000 UTC m=+958.099277901 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "metrics-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: I0314 09:26:29.988495 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.988716 4843 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:29 crc kubenswrapper[4843]: E0314 09:26:29.988791 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert podName:46ec1c5f-0334-4d5d-b16a-8e01ffc76613 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:31.988773821 +0000 UTC m=+959.301384949 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert") pod "infra-operator-controller-manager-54dc5b8f8d-ffxmw" (UID: "46ec1c5f-0334-4d5d-b16a-8e01ffc76613") : secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.100805 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" event={"ID":"0e6cdd9e-39ed-4359-9b74-54e23a0c70e2","Type":"ContainerStarted","Data":"34bff3ef1400911d1ddad07c87ccc647036c0d40e9198876f2133f03d86c46c8"} Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.103705 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" event={"ID":"09b1d07d-69b7-427b-9577-2a782637a2b1","Type":"ContainerStarted","Data":"1b21691a68f202c783069a64f22f32fcd4756038caa77524a12a1cc35a053f4f"} Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.295720 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.295931 4843 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.296005 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert podName:f5a4b439-9571-4fdb-8796-3bd729646aa3 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:32.295963371 +0000 UTC m=+959.608574499 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" (UID: "f5a4b439-9571-4fdb-8796-3bd729646aa3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.307714 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.325343 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7"] Mar 14 09:26:30 crc kubenswrapper[4843]: W0314 09:26:30.331760 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3da8acd2_64eb_4f3f_b28e_80d5bb4e4743.slice/crio-4a095cf740c0666e54291b88e32bf20b05d720c521bc44540476cafcaef0d12c WatchSource:0}: Error finding container 4a095cf740c0666e54291b88e32bf20b05d720c521bc44540476cafcaef0d12c: Status 404 returned error can't find the container with id 4a095cf740c0666e54291b88e32bf20b05d720c521bc44540476cafcaef0d12c Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.336266 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.344762 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-g59td"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.352319 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.361300 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh"] Mar 14 09:26:30 crc kubenswrapper[4843]: W0314 09:26:30.365468 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fbf4f22_f279_4841_b829_6d35d2dd0bfc.slice/crio-0289d22a296355f83a87884d1799f998045954e1bac5f51c6bbda62214beadb1 WatchSource:0}: Error finding container 0289d22a296355f83a87884d1799f998045954e1bac5f51c6bbda62214beadb1: Status 404 returned error can't find the container with id 0289d22a296355f83a87884d1799f998045954e1bac5f51c6bbda62214beadb1 Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.372956 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.394453 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.431003 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8"] Mar 14 09:26:30 crc kubenswrapper[4843]: W0314 09:26:30.445190 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod720b9936_b2c8_4b7b_ab8f_c8ecfda9052d.slice/crio-c60e2458205a88b7339812682d27e292226526e4536203dce409aa4480c6fb96 WatchSource:0}: Error finding container c60e2458205a88b7339812682d27e292226526e4536203dce409aa4480c6fb96: Status 404 returned error can't find the container with id c60e2458205a88b7339812682d27e292226526e4536203dce409aa4480c6fb96 Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.449871 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-czghc"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.456594 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm"] Mar 14 09:26:30 crc kubenswrapper[4843]: W0314 09:26:30.461453 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92eda9b1_e96f_4a71_93f8_a595c6f5117d.slice/crio-bb7b0e9671fd1a59693b1c8ce433a22daa3e9aa3430cac4f15b3e3fc43994ec9 WatchSource:0}: Error finding container bb7b0e9671fd1a59693b1c8ce433a22daa3e9aa3430cac4f15b3e3fc43994ec9: Status 404 returned error can't find the container with id bb7b0e9671fd1a59693b1c8ce433a22daa3e9aa3430cac4f15b3e3fc43994ec9 Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.461559 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr"] Mar 14 09:26:30 crc kubenswrapper[4843]: W0314 09:26:30.462027 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d4e28ff_8c1b_4533_a733_97c44a2b7b0a.slice/crio-c521f32a7eeddac40e87d0fedebb0f24c1ef63de99cd614e088ee023cc2a913a WatchSource:0}: Error finding container c521f32a7eeddac40e87d0fedebb0f24c1ef63de99cd614e088ee023cc2a913a: Status 404 returned error can't find the container with id c521f32a7eeddac40e87d0fedebb0f24c1ef63de99cd614e088ee023cc2a913a Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.468458 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:3a0fc90da4caf7412ae01e21542b53a10fe7a2732a705b0ae83f926d72c7332a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s75qh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6854b8b9d9-vgstg_openstack-operators(f1e0e04a-1194-4fa6-878f-318d360845d3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.469454 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddkk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-77b6666d85-czghc_openstack-operators(7d4e28ff-8c1b-4533-a733-97c44a2b7b0a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.469519 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" podUID="f1e0e04a-1194-4fa6-878f-318d360845d3" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.469697 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4274k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bbc5b68f9-xctxr_openstack-operators(92eda9b1-e96f-4a71-93f8-a595c6f5117d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.470073 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9ccmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-tx8dt_openstack-operators(bdefcc7b-402d-4c04-bc48-aa8b918d1b15): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.470893 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" podUID="92eda9b1-e96f-4a71-93f8-a595c6f5117d" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.470943 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" podUID="7d4e28ff-8c1b-4533-a733-97c44a2b7b0a" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.472049 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" podUID="bdefcc7b-402d-4c04-bc48-aa8b918d1b15" Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.473344 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c"] Mar 14 09:26:30 crc kubenswrapper[4843]: W0314 09:26:30.475305 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54b06835_eef3_4b4d_bf36_2714b438588f.slice/crio-9f64e496df6d89209f33d5b389a6ba33286d82ddab6d3d3ba19baac45942ee91 WatchSource:0}: Error finding container 9f64e496df6d89209f33d5b389a6ba33286d82ddab6d3d3ba19baac45942ee91: Status 404 returned error can't find the container with id 9f64e496df6d89209f33d5b389a6ba33286d82ddab6d3d3ba19baac45942ee91 Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.481537 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m7dz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5f4f55cb5c-zrfcm_openstack-operators(72b0bab9-36e7-49da-af7b-ff4fe9475545): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.481789 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-854cr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7b8d757b5d-2cr4c_openstack-operators(54b06835-eef3-4b4d-bf36-2714b438588f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.482693 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" podUID="72b0bab9-36e7-49da-af7b-ff4fe9475545" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.483007 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.484796 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5pmvh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-9nkkk_openstack-operators(99c2bbc7-1885-4031-9478-a839c36bfbc5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.485990 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" podUID="99c2bbc7-1885-4031-9478-a839c36bfbc5" Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.486033 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.494391 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.504774 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.509589 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.578381 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78"] Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.831185 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:30 crc kubenswrapper[4843]: I0314 09:26:30.831320 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.831399 4843 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.831480 4843 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.831487 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:32.8314624 +0000 UTC m=+960.144073528 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "webhook-server-cert" not found Mar 14 09:26:30 crc kubenswrapper[4843]: E0314 09:26:30.831558 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:32.831537282 +0000 UTC m=+960.144148520 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "metrics-server-cert" not found Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.115067 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" event={"ID":"92eda9b1-e96f-4a71-93f8-a595c6f5117d","Type":"ContainerStarted","Data":"bb7b0e9671fd1a59693b1c8ce433a22daa3e9aa3430cac4f15b3e3fc43994ec9"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.117890 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" event={"ID":"47f2d7ba-aac2-44f4-9146-ca20192a08ab","Type":"ContainerStarted","Data":"56a5aba1cc76f374a22d5b6df1e1a2552f8aa137a05e1102c9d08113a9b13f97"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.118497 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" podUID="92eda9b1-e96f-4a71-93f8-a595c6f5117d" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.119809 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" event={"ID":"0fbf4f22-f279-4841-b829-6d35d2dd0bfc","Type":"ContainerStarted","Data":"0289d22a296355f83a87884d1799f998045954e1bac5f51c6bbda62214beadb1"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.121622 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" event={"ID":"cf4074be-d697-47bf-a44c-57822afebb15","Type":"ContainerStarted","Data":"426fe765d24481051c03bff9baa2c70e7f71d7d442bbc1d4833cc88b2ed7904e"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.123940 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" event={"ID":"d4d69189-dfff-4f22-b978-4988c505df47","Type":"ContainerStarted","Data":"dc6a01c6a70f8e32d357102f80a72fff6c7c0804571a4b5d3dc199b403339d95"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.130888 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" event={"ID":"354f7562-fd32-491f-a3dd-b42bcf7d30cb","Type":"ContainerStarted","Data":"b34d35eed3f44b6daa9bba9fa3c877f6e3e814cf0c76dc15baf62e3dc6f27114"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.132335 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" event={"ID":"f75cfc62-a50d-4bad-9868-c58dc328cdbb","Type":"ContainerStarted","Data":"67990723c083fb953e6e5fd75680d66fff207925669450c33463078a80003d67"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.134415 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" event={"ID":"28f29a72-e196-4962-87b3-cc49b8591cc0","Type":"ContainerStarted","Data":"21bf8b59de48cbce5ac4ca94ab417e365979184b40b5dfcac2d65b498962e91a"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.135802 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" event={"ID":"ddd8e1af-7084-4685-96a9-ca4eaa9a4742","Type":"ContainerStarted","Data":"8678f6992986bc63cd14e0d86eb205e64d892db4fa85c16d835e229150c5c8c5"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.138649 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" event={"ID":"72b0bab9-36e7-49da-af7b-ff4fe9475545","Type":"ContainerStarted","Data":"8bd98582137b024eed00338dd90425fae0b17ebb1c194f7ee040f11d23a75b2c"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.142918 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" event={"ID":"3da8acd2-64eb-4f3f-b28e-80d5bb4e4743","Type":"ContainerStarted","Data":"4a095cf740c0666e54291b88e32bf20b05d720c521bc44540476cafcaef0d12c"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.161265 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" podUID="72b0bab9-36e7-49da-af7b-ff4fe9475545" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.165132 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" event={"ID":"54b06835-eef3-4b4d-bf36-2714b438588f","Type":"ContainerStarted","Data":"9f64e496df6d89209f33d5b389a6ba33286d82ddab6d3d3ba19baac45942ee91"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.170521 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.173557 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" event={"ID":"7d4e28ff-8c1b-4533-a733-97c44a2b7b0a","Type":"ContainerStarted","Data":"c521f32a7eeddac40e87d0fedebb0f24c1ef63de99cd614e088ee023cc2a913a"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.181745 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6\\\"\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" podUID="7d4e28ff-8c1b-4533-a733-97c44a2b7b0a" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.181941 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" event={"ID":"720b9936-b2c8-4b7b-ab8f-c8ecfda9052d","Type":"ContainerStarted","Data":"c60e2458205a88b7339812682d27e292226526e4536203dce409aa4480c6fb96"} Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.190867 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" event={"ID":"bdefcc7b-402d-4c04-bc48-aa8b918d1b15","Type":"ContainerStarted","Data":"ac7a369953c6e7c23e56a57e9cf955ffa9e82d044a009c7148c4803f9c683b3f"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.192252 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" podUID="bdefcc7b-402d-4c04-bc48-aa8b918d1b15" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.194917 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" event={"ID":"f1e0e04a-1194-4fa6-878f-318d360845d3","Type":"ContainerStarted","Data":"baedeb7ebb7d3eda6d8429ffdce9e1c1c258970606ba353abb12d3ff4878a4ca"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.199083 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:3a0fc90da4caf7412ae01e21542b53a10fe7a2732a705b0ae83f926d72c7332a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" podUID="f1e0e04a-1194-4fa6-878f-318d360845d3" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.204789 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" event={"ID":"99c2bbc7-1885-4031-9478-a839c36bfbc5","Type":"ContainerStarted","Data":"c002921c11cb62b15cb45574d2f46bbba0a8907994ff7a903b630bf2dd6efa7c"} Mar 14 09:26:31 crc kubenswrapper[4843]: E0314 09:26:31.207139 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" podUID="99c2bbc7-1885-4031-9478-a839c36bfbc5" Mar 14 09:26:31 crc kubenswrapper[4843]: I0314 09:26:31.208476 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" event={"ID":"abe14da3-e8a8-45f6-aebb-5df03c384a35","Type":"ContainerStarted","Data":"1d21a075dc57d3d0a117cd1c647efe4d97e00cfaeb330909c0dfd60d4ea10c4e"} Mar 14 09:26:32 crc kubenswrapper[4843]: I0314 09:26:32.053147 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.053348 4843 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.053470 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert podName:46ec1c5f-0334-4d5d-b16a-8e01ffc76613 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:36.053386035 +0000 UTC m=+963.365997223 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert") pod "infra-operator-controller-manager-54dc5b8f8d-ffxmw" (UID: "46ec1c5f-0334-4d5d-b16a-8e01ffc76613") : secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.224627 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" podUID="92eda9b1-e96f-4a71-93f8-a595c6f5117d" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.224745 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" podUID="72b0bab9-36e7-49da-af7b-ff4fe9475545" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.225049 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6\\\"\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" podUID="7d4e28ff-8c1b-4533-a733-97c44a2b7b0a" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.225145 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.225525 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:3a0fc90da4caf7412ae01e21542b53a10fe7a2732a705b0ae83f926d72c7332a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" podUID="f1e0e04a-1194-4fa6-878f-318d360845d3" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.225592 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" podUID="bdefcc7b-402d-4c04-bc48-aa8b918d1b15" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.225711 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" podUID="99c2bbc7-1885-4031-9478-a839c36bfbc5" Mar 14 09:26:32 crc kubenswrapper[4843]: I0314 09:26:32.364984 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.365134 4843 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.365182 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert podName:f5a4b439-9571-4fdb-8796-3bd729646aa3 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:36.365164822 +0000 UTC m=+963.677775950 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" (UID: "f5a4b439-9571-4fdb-8796-3bd729646aa3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: I0314 09:26:32.873328 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.873574 4843 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: I0314 09:26:32.873766 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.873837 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:36.873798956 +0000 UTC m=+964.186410084 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "webhook-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.873944 4843 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 14 09:26:32 crc kubenswrapper[4843]: E0314 09:26:32.874024 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:36.87400251 +0000 UTC m=+964.186613858 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "metrics-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: I0314 09:26:36.134983 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.136187 4843 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.136259 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert podName:46ec1c5f-0334-4d5d-b16a-8e01ffc76613 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:44.136236593 +0000 UTC m=+971.448847721 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert") pod "infra-operator-controller-manager-54dc5b8f8d-ffxmw" (UID: "46ec1c5f-0334-4d5d-b16a-8e01ffc76613") : secret "infra-operator-webhook-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: I0314 09:26:36.442836 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.443136 4843 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.443252 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert podName:f5a4b439-9571-4fdb-8796-3bd729646aa3 nodeName:}" failed. No retries permitted until 2026-03-14 09:26:44.443224087 +0000 UTC m=+971.755835215 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" (UID: "f5a4b439-9571-4fdb-8796-3bd729646aa3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: I0314 09:26:36.950740 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:36 crc kubenswrapper[4843]: I0314 09:26:36.950834 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.950926 4843 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.951001 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:44.950983039 +0000 UTC m=+972.263594167 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "webhook-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.951023 4843 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 14 09:26:36 crc kubenswrapper[4843]: E0314 09:26:36.951082 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:26:44.951064781 +0000 UTC m=+972.263675999 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "metrics-server-cert" not found Mar 14 09:26:40 crc kubenswrapper[4843]: I0314 09:26:40.587827 4843 scope.go:117] "RemoveContainer" containerID="46385c8fe57811cadbd73b503074c929b631e8a732af4004508a58abc3c155f7" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.189405 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.196394 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46ec1c5f-0334-4d5d-b16a-8e01ffc76613-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-ffxmw\" (UID: \"46ec1c5f-0334-4d5d-b16a-8e01ffc76613\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:44 crc kubenswrapper[4843]: E0314 09:26:44.440465 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca" Mar 14 09:26:44 crc kubenswrapper[4843]: E0314 09:26:44.440644 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6l9fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-684f77d66d-7v8h2_openstack-operators(d4d69189-dfff-4f22-b978-4988c505df47): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:26:44 crc kubenswrapper[4843]: E0314 09:26:44.441894 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" podUID="d4d69189-dfff-4f22-b978-4988c505df47" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.486234 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4jsrl" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.494324 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.495076 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.499950 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5a4b439-9571-4fdb-8796-3bd729646aa3-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7slhjt\" (UID: \"f5a4b439-9571-4fdb-8796-3bd729646aa3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.528262 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-88shv" Mar 14 09:26:44 crc kubenswrapper[4843]: I0314 09:26:44.535796 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:26:44 crc kubenswrapper[4843]: E0314 09:26:44.895159 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Mar 14 09:26:44 crc kubenswrapper[4843]: E0314 09:26:44.895372 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c4nbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qjs78_openstack-operators(f75cfc62-a50d-4bad-9868-c58dc328cdbb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:26:44 crc kubenswrapper[4843]: E0314 09:26:44.896786 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" podUID="f75cfc62-a50d-4bad-9868-c58dc328cdbb" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.002157 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:45 crc kubenswrapper[4843]: E0314 09:26:45.002192 4843 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 14 09:26:45 crc kubenswrapper[4843]: E0314 09:26:45.002259 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:27:01.002244702 +0000 UTC m=+988.314855830 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "webhook-server-cert" not found Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.002330 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:26:45 crc kubenswrapper[4843]: E0314 09:26:45.002462 4843 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 14 09:26:45 crc kubenswrapper[4843]: E0314 09:26:45.002506 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs podName:a0339dbb-a75e-465b-a30a-75c3091f506c nodeName:}" failed. No retries permitted until 2026-03-14 09:27:01.002495008 +0000 UTC m=+988.315106136 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs") pod "openstack-operator-controller-manager-59cdb7596d-l4ggb" (UID: "a0339dbb-a75e-465b-a30a-75c3091f506c") : secret "metrics-server-cert" not found Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.329939 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" event={"ID":"0e6cdd9e-39ed-4359-9b74-54e23a0c70e2","Type":"ContainerStarted","Data":"1382ce47b11d88ffbccd0145ef17fef5e0c55ce96521d2cce9fdeddb1c770d97"} Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.330370 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.332832 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" event={"ID":"abe14da3-e8a8-45f6-aebb-5df03c384a35","Type":"ContainerStarted","Data":"3ecce86ef5ee4963bcd3efaffea599e9ee1f4f385d8e045b9a8ba7616b343164"} Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.333379 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.345199 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" podStartSLOduration=2.044330547 podStartE2EDuration="17.345184029s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:29.62369385 +0000 UTC m=+956.936304978" lastFinishedPulling="2026-03-14 09:26:44.924547332 +0000 UTC m=+972.237158460" observedRunningTime="2026-03-14 09:26:45.344206745 +0000 UTC m=+972.656817873" watchObservedRunningTime="2026-03-14 09:26:45.345184029 +0000 UTC m=+972.657795147" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.365252 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" event={"ID":"354f7562-fd32-491f-a3dd-b42bcf7d30cb","Type":"ContainerStarted","Data":"d82f1b4fd2e6add4511855a4b31981e6d07870cef2175e01ff61d87e69d46799"} Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.365332 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" event={"ID":"09b1d07d-69b7-427b-9577-2a782637a2b1","Type":"ContainerStarted","Data":"753ac91e2571aca946f963f0f716d1acf94d708edece589d24c5d2cea51a92e2"} Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.365387 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.365407 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.371314 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" podStartSLOduration=2.8928127139999997 podStartE2EDuration="17.371299018s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.445151146 +0000 UTC m=+957.757762274" lastFinishedPulling="2026-03-14 09:26:44.92363745 +0000 UTC m=+972.236248578" observedRunningTime="2026-03-14 09:26:45.363511773 +0000 UTC m=+972.676122891" watchObservedRunningTime="2026-03-14 09:26:45.371299018 +0000 UTC m=+972.683910146" Mar 14 09:26:45 crc kubenswrapper[4843]: E0314 09:26:45.372026 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" podUID="f75cfc62-a50d-4bad-9868-c58dc328cdbb" Mar 14 09:26:45 crc kubenswrapper[4843]: E0314 09:26:45.372053 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" podUID="d4d69189-dfff-4f22-b978-4988c505df47" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.396677 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.396738 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.451863 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" podStartSLOduration=2.9648636010000002 podStartE2EDuration="17.451838606s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.438231572 +0000 UTC m=+957.750842700" lastFinishedPulling="2026-03-14 09:26:44.925206577 +0000 UTC m=+972.237817705" observedRunningTime="2026-03-14 09:26:45.440013066 +0000 UTC m=+972.752624194" watchObservedRunningTime="2026-03-14 09:26:45.451838606 +0000 UTC m=+972.764449744" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.461366 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" podStartSLOduration=10.506821836 podStartE2EDuration="17.461345991s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:29.476936543 +0000 UTC m=+956.789547671" lastFinishedPulling="2026-03-14 09:26:36.431460698 +0000 UTC m=+963.744071826" observedRunningTime="2026-03-14 09:26:45.459540239 +0000 UTC m=+972.772151357" watchObservedRunningTime="2026-03-14 09:26:45.461345991 +0000 UTC m=+972.773957119" Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.471829 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw"] Mar 14 09:26:45 crc kubenswrapper[4843]: W0314 09:26:45.475256 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46ec1c5f_0334_4d5d_b16a_8e01ffc76613.slice/crio-74b2c370cb77c14ad89b73f383db2d52a7a4524b6f40446c8170dd7105fba8be WatchSource:0}: Error finding container 74b2c370cb77c14ad89b73f383db2d52a7a4524b6f40446c8170dd7105fba8be: Status 404 returned error can't find the container with id 74b2c370cb77c14ad89b73f383db2d52a7a4524b6f40446c8170dd7105fba8be Mar 14 09:26:45 crc kubenswrapper[4843]: I0314 09:26:45.569986 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt"] Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.367259 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" event={"ID":"46ec1c5f-0334-4d5d-b16a-8e01ffc76613","Type":"ContainerStarted","Data":"74b2c370cb77c14ad89b73f383db2d52a7a4524b6f40446c8170dd7105fba8be"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.370971 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" event={"ID":"28f29a72-e196-4962-87b3-cc49b8591cc0","Type":"ContainerStarted","Data":"040ce80e921105039cab16e51882ae8f7ef54ad6d5ecd86ade0bed2449c21527"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.371840 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.392779 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" podStartSLOduration=3.93038893 podStartE2EDuration="18.392763552s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.461662017 +0000 UTC m=+957.774273145" lastFinishedPulling="2026-03-14 09:26:44.924036639 +0000 UTC m=+972.236647767" observedRunningTime="2026-03-14 09:26:46.391075912 +0000 UTC m=+973.703687040" watchObservedRunningTime="2026-03-14 09:26:46.392763552 +0000 UTC m=+973.705374680" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.427972 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" event={"ID":"47f2d7ba-aac2-44f4-9146-ca20192a08ab","Type":"ContainerStarted","Data":"a073f5110e6f1ea6b744a5a5a41be85bebc72a78ef2bd0fea48a5cd16fb05350"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.428043 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.435017 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" event={"ID":"0fbf4f22-f279-4841-b829-6d35d2dd0bfc","Type":"ContainerStarted","Data":"9248d1366fbc7dccfee16c11c3b6a8e8d097bbc525854985ccdefb027afe6075"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.435769 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.440005 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" event={"ID":"ddd8e1af-7084-4685-96a9-ca4eaa9a4742","Type":"ContainerStarted","Data":"c8b835a05a417522b499e91369150e356d1f4fa9c700227223daf747a470e6ef"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.440717 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.442252 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" podStartSLOduration=3.938971342 podStartE2EDuration="18.442239734s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.421802392 +0000 UTC m=+957.734413520" lastFinishedPulling="2026-03-14 09:26:44.925070784 +0000 UTC m=+972.237681912" observedRunningTime="2026-03-14 09:26:46.439336965 +0000 UTC m=+973.751948113" watchObservedRunningTime="2026-03-14 09:26:46.442239734 +0000 UTC m=+973.754850862" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.454244 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" event={"ID":"720b9936-b2c8-4b7b-ab8f-c8ecfda9052d","Type":"ContainerStarted","Data":"e980fd87cf78a782290fadc9105a11ded8e690c01a420db477726d65ad8e774e"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.454970 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.456917 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" event={"ID":"f5a4b439-9571-4fdb-8796-3bd729646aa3","Type":"ContainerStarted","Data":"6c87e792b5cd9ed3b98f845d9a2af41665d747d224736f58b55d2918ae87ecab"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.466765 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" event={"ID":"3da8acd2-64eb-4f3f-b28e-80d5bb4e4743","Type":"ContainerStarted","Data":"054a976e5a0600d0d10dd65d96e0d523d5ada15bb885fb1e1d24822c6d703d1a"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.466873 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.469420 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" podStartSLOduration=3.925023933 podStartE2EDuration="18.469400728s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.378820564 +0000 UTC m=+957.691431692" lastFinishedPulling="2026-03-14 09:26:44.923197359 +0000 UTC m=+972.235808487" observedRunningTime="2026-03-14 09:26:46.462670788 +0000 UTC m=+973.775281926" watchObservedRunningTime="2026-03-14 09:26:46.469400728 +0000 UTC m=+973.782011856" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.469651 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" event={"ID":"cf4074be-d697-47bf-a44c-57822afebb15","Type":"ContainerStarted","Data":"9d21c3db70387e8c98581f2a1ad124e86afd18806869fd81bfee4592b225402d"} Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.470155 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.517486 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" podStartSLOduration=4.05529146 podStartE2EDuration="18.517466007s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.461439842 +0000 UTC m=+957.774050980" lastFinishedPulling="2026-03-14 09:26:44.923614399 +0000 UTC m=+972.236225527" observedRunningTime="2026-03-14 09:26:46.492633368 +0000 UTC m=+973.805244516" watchObservedRunningTime="2026-03-14 09:26:46.517466007 +0000 UTC m=+973.830077135" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.518353 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" podStartSLOduration=4.056240343 podStartE2EDuration="18.518346238s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.462092177 +0000 UTC m=+957.774703305" lastFinishedPulling="2026-03-14 09:26:44.924198072 +0000 UTC m=+972.236809200" observedRunningTime="2026-03-14 09:26:46.514281252 +0000 UTC m=+973.826892400" watchObservedRunningTime="2026-03-14 09:26:46.518346238 +0000 UTC m=+973.830957366" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.544754 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" podStartSLOduration=4.053692463 podStartE2EDuration="18.544736423s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.433095801 +0000 UTC m=+957.745706929" lastFinishedPulling="2026-03-14 09:26:44.924139761 +0000 UTC m=+972.236750889" observedRunningTime="2026-03-14 09:26:46.538672549 +0000 UTC m=+973.851283677" watchObservedRunningTime="2026-03-14 09:26:46.544736423 +0000 UTC m=+973.857347551" Mar 14 09:26:46 crc kubenswrapper[4843]: I0314 09:26:46.570659 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" podStartSLOduration=4.116139382 podStartE2EDuration="18.570635207s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.468029738 +0000 UTC m=+957.780640866" lastFinishedPulling="2026-03-14 09:26:44.922525563 +0000 UTC m=+972.235136691" observedRunningTime="2026-03-14 09:26:46.558553741 +0000 UTC m=+973.871164869" watchObservedRunningTime="2026-03-14 09:26:46.570635207 +0000 UTC m=+973.883246335" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.425895 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-d47688694-ktr4z" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.447469 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-75kt6" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.461066 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-mxn5j" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.490772 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-g29z7" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.563007 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-bmb27" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.608413 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-8wvhp" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.832855 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-fqfbh" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.841852 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.905745 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-9dlkn" Mar 14 09:26:58 crc kubenswrapper[4843]: I0314 09:26:58.906125 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7f84474648-g59td" Mar 14 09:26:59 crc kubenswrapper[4843]: I0314 09:26:59.005856 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-kbcdm" Mar 14 09:27:01 crc kubenswrapper[4843]: I0314 09:27:01.010069 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:01 crc kubenswrapper[4843]: I0314 09:27:01.010154 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:01 crc kubenswrapper[4843]: I0314 09:27:01.015828 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-metrics-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:01 crc kubenswrapper[4843]: I0314 09:27:01.031617 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a0339dbb-a75e-465b-a30a-75c3091f506c-webhook-certs\") pod \"openstack-operator-controller-manager-59cdb7596d-l4ggb\" (UID: \"a0339dbb-a75e-465b-a30a-75c3091f506c\") " pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:01 crc kubenswrapper[4843]: I0314 09:27:01.210022 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9t5km" Mar 14 09:27:01 crc kubenswrapper[4843]: I0314 09:27:01.218507 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:04 crc kubenswrapper[4843]: E0314 09:27:04.976852 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:3a0fc90da4caf7412ae01e21542b53a10fe7a2732a705b0ae83f926d72c7332a" Mar 14 09:27:04 crc kubenswrapper[4843]: E0314 09:27:04.977281 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:3a0fc90da4caf7412ae01e21542b53a10fe7a2732a705b0ae83f926d72c7332a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s75qh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6854b8b9d9-vgstg_openstack-operators(f1e0e04a-1194-4fa6-878f-318d360845d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:27:04 crc kubenswrapper[4843]: E0314 09:27:04.978525 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" podUID="f1e0e04a-1194-4fa6-878f-318d360845d3" Mar 14 09:27:04 crc kubenswrapper[4843]: E0314 09:27:04.979948 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978" Mar 14 09:27:04 crc kubenswrapper[4843]: E0314 09:27:04.980230 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9ccmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-tx8dt_openstack-operators(bdefcc7b-402d-4c04-bc48-aa8b918d1b15): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:27:04 crc kubenswrapper[4843]: E0314 09:27:04.981772 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" podUID="bdefcc7b-402d-4c04-bc48-aa8b918d1b15" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.335460 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.335897 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddkk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-77b6666d85-czghc_openstack-operators(7d4e28ff-8c1b-4533-a733-97c44a2b7b0a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.337033 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" podUID="7d4e28ff-8c1b-4533-a733-97c44a2b7b0a" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.442724 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.442798 4843 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.443048 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-854cr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7b8d757b5d-2cr4c_openstack-operators(54b06835-eef3-4b4d-bf36-2714b438588f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:27:05 crc kubenswrapper[4843]: E0314 09:27:05.444265 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.010471 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb"] Mar 14 09:27:06 crc kubenswrapper[4843]: W0314 09:27:06.020198 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0339dbb_a75e_465b_a30a_75c3091f506c.slice/crio-7f81a8f9eeb511c9a8ce73fc25b2dccefbe4be57cad97ebb58bc11b80794829a WatchSource:0}: Error finding container 7f81a8f9eeb511c9a8ce73fc25b2dccefbe4be57cad97ebb58bc11b80794829a: Status 404 returned error can't find the container with id 7f81a8f9eeb511c9a8ce73fc25b2dccefbe4be57cad97ebb58bc11b80794829a Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.640902 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" event={"ID":"72b0bab9-36e7-49da-af7b-ff4fe9475545","Type":"ContainerStarted","Data":"d94acf7199fbb393d498a3b90aefe190fc573d7b1239ea770aab384ad29f5039"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.641948 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.642964 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" event={"ID":"f5a4b439-9571-4fdb-8796-3bd729646aa3","Type":"ContainerStarted","Data":"dbfc49e3efde7deae6ff9f2171c18725aa1ca7020c93937b4faf33c881ed5495"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.643023 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.644177 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" event={"ID":"d4d69189-dfff-4f22-b978-4988c505df47","Type":"ContainerStarted","Data":"67d7fb422b668e3e484c29c48709bc2437bf101288d7e225f0257dbed9ea380e"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.644382 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.645557 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" event={"ID":"46ec1c5f-0334-4d5d-b16a-8e01ffc76613","Type":"ContainerStarted","Data":"014330a90c38ea704bcaf39e1cc0c120c825c5a34b7954de49f5479e8082b6ac"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.645664 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.647353 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" event={"ID":"99c2bbc7-1885-4031-9478-a839c36bfbc5","Type":"ContainerStarted","Data":"3a603f6481416c1877cd5aec8ed67e2646ca944c4dc0cd7db4cf2ab6b70ac209"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.647792 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.649083 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" event={"ID":"92eda9b1-e96f-4a71-93f8-a595c6f5117d","Type":"ContainerStarted","Data":"733b7bece610d210e61ace229e757beda71b5471f92abce3cda81229478208f5"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.649471 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.650863 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" event={"ID":"a0339dbb-a75e-465b-a30a-75c3091f506c","Type":"ContainerStarted","Data":"1d81c978ae73682faff34f87bc7b8a60008fbb2cbb03b208d177419fe6ec6994"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.651091 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" event={"ID":"a0339dbb-a75e-465b-a30a-75c3091f506c","Type":"ContainerStarted","Data":"7f81a8f9eeb511c9a8ce73fc25b2dccefbe4be57cad97ebb58bc11b80794829a"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.651142 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.652418 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" event={"ID":"f75cfc62-a50d-4bad-9868-c58dc328cdbb","Type":"ContainerStarted","Data":"cc16fdc933c6864cf2ec73462ad4551ed3df60de73b45fabcf79d2a32905743f"} Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.664112 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" podStartSLOduration=3.61584682 podStartE2EDuration="38.664089408s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.481400525 +0000 UTC m=+957.794011653" lastFinishedPulling="2026-03-14 09:27:05.529643103 +0000 UTC m=+992.842254241" observedRunningTime="2026-03-14 09:27:06.658160276 +0000 UTC m=+993.970771404" watchObservedRunningTime="2026-03-14 09:27:06.664089408 +0000 UTC m=+993.976700536" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.680730 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjs78" podStartSLOduration=3.704471304 podStartE2EDuration="38.680698756s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.586900694 +0000 UTC m=+957.899511822" lastFinishedPulling="2026-03-14 09:27:05.563128146 +0000 UTC m=+992.875739274" observedRunningTime="2026-03-14 09:27:06.672020088 +0000 UTC m=+993.984631216" watchObservedRunningTime="2026-03-14 09:27:06.680698756 +0000 UTC m=+993.993309884" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.704380 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" podStartSLOduration=18.730278826 podStartE2EDuration="38.704359304s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:45.588171376 +0000 UTC m=+972.900782504" lastFinishedPulling="2026-03-14 09:27:05.562251854 +0000 UTC m=+992.874862982" observedRunningTime="2026-03-14 09:27:06.70375856 +0000 UTC m=+994.016369688" watchObservedRunningTime="2026-03-14 09:27:06.704359304 +0000 UTC m=+994.016970432" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.730154 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" podStartSLOduration=3.629699031 podStartE2EDuration="38.730136142s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.484720693 +0000 UTC m=+957.797331821" lastFinishedPulling="2026-03-14 09:27:05.585157804 +0000 UTC m=+992.897768932" observedRunningTime="2026-03-14 09:27:06.726105456 +0000 UTC m=+994.038716584" watchObservedRunningTime="2026-03-14 09:27:06.730136142 +0000 UTC m=+994.042747270" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.784252 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" podStartSLOduration=38.784230479 podStartE2EDuration="38.784230479s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:27:06.780493139 +0000 UTC m=+994.093104267" watchObservedRunningTime="2026-03-14 09:27:06.784230479 +0000 UTC m=+994.096841607" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.812028 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" podStartSLOduration=3.719318247 podStartE2EDuration="38.812007676s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.469511604 +0000 UTC m=+957.782122742" lastFinishedPulling="2026-03-14 09:27:05.562201043 +0000 UTC m=+992.874812171" observedRunningTime="2026-03-14 09:27:06.804833884 +0000 UTC m=+994.117445012" watchObservedRunningTime="2026-03-14 09:27:06.812007676 +0000 UTC m=+994.124618804" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.835814 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" podStartSLOduration=3.734725386 podStartE2EDuration="38.835796876s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.462051236 +0000 UTC m=+957.774662364" lastFinishedPulling="2026-03-14 09:27:05.563122726 +0000 UTC m=+992.875733854" observedRunningTime="2026-03-14 09:27:06.822675361 +0000 UTC m=+994.135286499" watchObservedRunningTime="2026-03-14 09:27:06.835796876 +0000 UTC m=+994.148407994" Mar 14 09:27:06 crc kubenswrapper[4843]: I0314 09:27:06.849817 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" podStartSLOduration=18.781858 podStartE2EDuration="38.849800572s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:45.494460766 +0000 UTC m=+972.807071894" lastFinishedPulling="2026-03-14 09:27:05.562403328 +0000 UTC m=+992.875014466" observedRunningTime="2026-03-14 09:27:06.84803479 +0000 UTC m=+994.160645928" watchObservedRunningTime="2026-03-14 09:27:06.849800572 +0000 UTC m=+994.162411720" Mar 14 09:27:11 crc kubenswrapper[4843]: I0314 09:27:11.226977 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-59cdb7596d-l4ggb" Mar 14 09:27:14 crc kubenswrapper[4843]: I0314 09:27:14.499909 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-ffxmw" Mar 14 09:27:14 crc kubenswrapper[4843]: I0314 09:27:14.545472 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7slhjt" Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.387711 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.388161 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.388483 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.391554 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"144fcfbfce69a65758af21b0b25db69c4308200478b4f5a2c753f3696a587b2b"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.392464 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://144fcfbfce69a65758af21b0b25db69c4308200478b4f5a2c753f3696a587b2b" gracePeriod=600 Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.723768 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="144fcfbfce69a65758af21b0b25db69c4308200478b4f5a2c753f3696a587b2b" exitCode=0 Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.723812 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"144fcfbfce69a65758af21b0b25db69c4308200478b4f5a2c753f3696a587b2b"} Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.724064 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"1c0b6bcfd5822f50fc94ed067fd20658758c736e480b5fd0ec55da8213db2e74"} Mar 14 09:27:15 crc kubenswrapper[4843]: I0314 09:27:15.724081 4843 scope.go:117] "RemoveContainer" containerID="0419057bafaf42ee3e8fe2dcca9e3ad2340132e208a0d8b9f235557b85f11481" Mar 14 09:27:16 crc kubenswrapper[4843]: E0314 09:27:16.340429 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6\\\"\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" podUID="7d4e28ff-8c1b-4533-a733-97c44a2b7b0a" Mar 14 09:27:16 crc kubenswrapper[4843]: E0314 09:27:16.340657 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:3a0fc90da4caf7412ae01e21542b53a10fe7a2732a705b0ae83f926d72c7332a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" podUID="f1e0e04a-1194-4fa6-878f-318d360845d3" Mar 14 09:27:17 crc kubenswrapper[4843]: E0314 09:27:17.341074 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" podUID="bdefcc7b-402d-4c04-bc48-aa8b918d1b15" Mar 14 09:27:18 crc kubenswrapper[4843]: E0314 09:27:18.340557 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.243:5001/openstack-k8s-operators/watcher-operator:0d9467a022a6d6a8b6465cdbc9e4aefc533e796a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" Mar 14 09:27:18 crc kubenswrapper[4843]: I0314 09:27:18.723555 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7v8h2" Mar 14 09:27:18 crc kubenswrapper[4843]: I0314 09:27:18.911565 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-zrfcm" Mar 14 09:27:18 crc kubenswrapper[4843]: I0314 09:27:18.958467 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xctxr" Mar 14 09:27:19 crc kubenswrapper[4843]: I0314 09:27:19.074844 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-9nkkk" Mar 14 09:27:29 crc kubenswrapper[4843]: I0314 09:27:29.835451 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" event={"ID":"bdefcc7b-402d-4c04-bc48-aa8b918d1b15","Type":"ContainerStarted","Data":"97eb3be71406e96122bee0ff280bdc760e3f716affd075905ed78e377e4521e7"} Mar 14 09:27:29 crc kubenswrapper[4843]: I0314 09:27:29.836175 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:27:29 crc kubenswrapper[4843]: I0314 09:27:29.837336 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" event={"ID":"f1e0e04a-1194-4fa6-878f-318d360845d3","Type":"ContainerStarted","Data":"65e0fe1ada515fb835a9eccfb905e05ee61d158f2a63633b88bf4b227afdc1a2"} Mar 14 09:27:29 crc kubenswrapper[4843]: I0314 09:27:29.837518 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:27:29 crc kubenswrapper[4843]: I0314 09:27:29.851685 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" podStartSLOduration=3.525813595 podStartE2EDuration="1m1.851659881s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.470007215 +0000 UTC m=+957.782618343" lastFinishedPulling="2026-03-14 09:27:28.795853501 +0000 UTC m=+1016.108464629" observedRunningTime="2026-03-14 09:27:29.848521476 +0000 UTC m=+1017.161132604" watchObservedRunningTime="2026-03-14 09:27:29.851659881 +0000 UTC m=+1017.164271009" Mar 14 09:27:31 crc kubenswrapper[4843]: I0314 09:27:31.361253 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" podStartSLOduration=5.036208727 podStartE2EDuration="1m3.361225002s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.468332176 +0000 UTC m=+957.780943294" lastFinishedPulling="2026-03-14 09:27:28.793348431 +0000 UTC m=+1016.105959569" observedRunningTime="2026-03-14 09:27:29.871538167 +0000 UTC m=+1017.184149285" watchObservedRunningTime="2026-03-14 09:27:31.361225002 +0000 UTC m=+1018.673836130" Mar 14 09:27:32 crc kubenswrapper[4843]: I0314 09:27:32.859019 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" event={"ID":"7d4e28ff-8c1b-4533-a733-97c44a2b7b0a","Type":"ContainerStarted","Data":"f3457d716143162058315e7ede0d0bd52983ea8d494e1ce0eff1dd83a67231d2"} Mar 14 09:27:32 crc kubenswrapper[4843]: I0314 09:27:32.859634 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:27:32 crc kubenswrapper[4843]: I0314 09:27:32.883205 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" podStartSLOduration=3.56855711 podStartE2EDuration="1m4.883186421s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.469351219 +0000 UTC m=+957.781962347" lastFinishedPulling="2026-03-14 09:27:31.78398053 +0000 UTC m=+1019.096591658" observedRunningTime="2026-03-14 09:27:32.87810436 +0000 UTC m=+1020.190715518" watchObservedRunningTime="2026-03-14 09:27:32.883186421 +0000 UTC m=+1020.195797549" Mar 14 09:27:33 crc kubenswrapper[4843]: I0314 09:27:33.869981 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" event={"ID":"54b06835-eef3-4b4d-bf36-2714b438588f","Type":"ContainerStarted","Data":"004cdb8e091a09caf9253b3da5bc483cdb68a90db98cd218b7e11f75b13f2e61"} Mar 14 09:27:33 crc kubenswrapper[4843]: I0314 09:27:33.909362 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podStartSLOduration=2.966478827 podStartE2EDuration="1m5.909329429s" podCreationTimestamp="2026-03-14 09:26:28 +0000 UTC" firstStartedPulling="2026-03-14 09:26:30.48164384 +0000 UTC m=+957.794254968" lastFinishedPulling="2026-03-14 09:27:33.424494442 +0000 UTC m=+1020.737105570" observedRunningTime="2026-03-14 09:27:33.89641426 +0000 UTC m=+1021.209025408" watchObservedRunningTime="2026-03-14 09:27:33.909329429 +0000 UTC m=+1021.221940557" Mar 14 09:27:38 crc kubenswrapper[4843]: I0314 09:27:38.501775 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-czghc" Mar 14 09:27:38 crc kubenswrapper[4843]: I0314 09:27:38.946891 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tx8dt" Mar 14 09:27:38 crc kubenswrapper[4843]: I0314 09:27:38.968250 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-vgstg" Mar 14 09:27:39 crc kubenswrapper[4843]: I0314 09:27:39.136933 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:27:39 crc kubenswrapper[4843]: I0314 09:27:39.139319 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.379501 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c"] Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.381263 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" containerName="manager" containerID="cri-o://004cdb8e091a09caf9253b3da5bc483cdb68a90db98cd218b7e11f75b13f2e61" gracePeriod=10 Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.427614 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp"] Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.427871 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" podUID="af58a715-edeb-480c-98a1-f42077825287" containerName="operator" containerID="cri-o://da3c83f872d4062855f44c6aa2a2ecdda7254b497fb9af510c75abca102cc7fe" gracePeriod=10 Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.971178 4843 generic.go:334] "Generic (PLEG): container finished" podID="54b06835-eef3-4b4d-bf36-2714b438588f" containerID="004cdb8e091a09caf9253b3da5bc483cdb68a90db98cd218b7e11f75b13f2e61" exitCode=0 Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.971569 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" event={"ID":"54b06835-eef3-4b4d-bf36-2714b438588f","Type":"ContainerDied","Data":"004cdb8e091a09caf9253b3da5bc483cdb68a90db98cd218b7e11f75b13f2e61"} Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.973333 4843 generic.go:334] "Generic (PLEG): container finished" podID="af58a715-edeb-480c-98a1-f42077825287" containerID="da3c83f872d4062855f44c6aa2a2ecdda7254b497fb9af510c75abca102cc7fe" exitCode=0 Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.973361 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" event={"ID":"af58a715-edeb-480c-98a1-f42077825287","Type":"ContainerDied","Data":"da3c83f872d4062855f44c6aa2a2ecdda7254b497fb9af510c75abca102cc7fe"} Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.973376 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" event={"ID":"af58a715-edeb-480c-98a1-f42077825287","Type":"ContainerDied","Data":"115af8b17df258c7634dc8eb91aa7b4cf2d3041b17bf3ccefb1863485418a90c"} Mar 14 09:27:45 crc kubenswrapper[4843]: I0314 09:27:45.973386 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="115af8b17df258c7634dc8eb91aa7b4cf2d3041b17bf3ccefb1863485418a90c" Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.038222 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.045127 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.204914 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-854cr\" (UniqueName: \"kubernetes.io/projected/54b06835-eef3-4b4d-bf36-2714b438588f-kube-api-access-854cr\") pod \"54b06835-eef3-4b4d-bf36-2714b438588f\" (UID: \"54b06835-eef3-4b4d-bf36-2714b438588f\") " Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.204962 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2qfb\" (UniqueName: \"kubernetes.io/projected/af58a715-edeb-480c-98a1-f42077825287-kube-api-access-w2qfb\") pod \"af58a715-edeb-480c-98a1-f42077825287\" (UID: \"af58a715-edeb-480c-98a1-f42077825287\") " Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.212664 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af58a715-edeb-480c-98a1-f42077825287-kube-api-access-w2qfb" (OuterVolumeSpecName: "kube-api-access-w2qfb") pod "af58a715-edeb-480c-98a1-f42077825287" (UID: "af58a715-edeb-480c-98a1-f42077825287"). InnerVolumeSpecName "kube-api-access-w2qfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.212761 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54b06835-eef3-4b4d-bf36-2714b438588f-kube-api-access-854cr" (OuterVolumeSpecName: "kube-api-access-854cr") pod "54b06835-eef3-4b4d-bf36-2714b438588f" (UID: "54b06835-eef3-4b4d-bf36-2714b438588f"). InnerVolumeSpecName "kube-api-access-854cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.306416 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-854cr\" (UniqueName: \"kubernetes.io/projected/54b06835-eef3-4b4d-bf36-2714b438588f-kube-api-access-854cr\") on node \"crc\" DevicePath \"\"" Mar 14 09:27:46 crc kubenswrapper[4843]: I0314 09:27:46.306452 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2qfb\" (UniqueName: \"kubernetes.io/projected/af58a715-edeb-480c-98a1-f42077825287-kube-api-access-w2qfb\") on node \"crc\" DevicePath \"\"" Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.014984 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp" Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.014991 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" event={"ID":"54b06835-eef3-4b4d-bf36-2714b438588f","Type":"ContainerDied","Data":"9f64e496df6d89209f33d5b389a6ba33286d82ddab6d3d3ba19baac45942ee91"} Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.015047 4843 scope.go:117] "RemoveContainer" containerID="004cdb8e091a09caf9253b3da5bc483cdb68a90db98cd218b7e11f75b13f2e61" Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.014992 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c" Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.068567 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp"] Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.075875 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bdb46c895-cstpp"] Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.085714 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c"] Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.091987 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7b8d757b5d-2cr4c"] Mar 14 09:27:47 crc kubenswrapper[4843]: E0314 09:27:47.094969 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54b06835_eef3_4b4d_bf36_2714b438588f.slice\": RecentStats: unable to find data in memory cache]" Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.350500 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" path="/var/lib/kubelet/pods/54b06835-eef3-4b4d-bf36-2714b438588f/volumes" Mar 14 09:27:47 crc kubenswrapper[4843]: I0314 09:27:47.351806 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af58a715-edeb-480c-98a1-f42077825287" path="/var/lib/kubelet/pods/af58a715-edeb-480c-98a1-f42077825287/volumes" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.577737 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-index-2ctxw"] Mar 14 09:27:50 crc kubenswrapper[4843]: E0314 09:27:50.578883 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" containerName="manager" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.578908 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" containerName="manager" Mar 14 09:27:50 crc kubenswrapper[4843]: E0314 09:27:50.578938 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af58a715-edeb-480c-98a1-f42077825287" containerName="operator" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.578950 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="af58a715-edeb-480c-98a1-f42077825287" containerName="operator" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.579183 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="af58a715-edeb-480c-98a1-f42077825287" containerName="operator" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.579222 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="54b06835-eef3-4b4d-bf36-2714b438588f" containerName="manager" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.580148 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.582910 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-index-dockercfg-8c9sw" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.588701 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-index-2ctxw"] Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.671967 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hsd9\" (UniqueName: \"kubernetes.io/projected/db24d096-30dd-4099-82e8-fe87d671eb7d-kube-api-access-6hsd9\") pod \"watcher-operator-index-2ctxw\" (UID: \"db24d096-30dd-4099-82e8-fe87d671eb7d\") " pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.773309 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hsd9\" (UniqueName: \"kubernetes.io/projected/db24d096-30dd-4099-82e8-fe87d671eb7d-kube-api-access-6hsd9\") pod \"watcher-operator-index-2ctxw\" (UID: \"db24d096-30dd-4099-82e8-fe87d671eb7d\") " pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.835370 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hsd9\" (UniqueName: \"kubernetes.io/projected/db24d096-30dd-4099-82e8-fe87d671eb7d-kube-api-access-6hsd9\") pod \"watcher-operator-index-2ctxw\" (UID: \"db24d096-30dd-4099-82e8-fe87d671eb7d\") " pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:50 crc kubenswrapper[4843]: I0314 09:27:50.901693 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:51 crc kubenswrapper[4843]: I0314 09:27:51.424480 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-index-2ctxw"] Mar 14 09:27:52 crc kubenswrapper[4843]: I0314 09:27:52.054864 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-2ctxw" event={"ID":"db24d096-30dd-4099-82e8-fe87d671eb7d","Type":"ContainerStarted","Data":"658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553"} Mar 14 09:27:52 crc kubenswrapper[4843]: I0314 09:27:52.055190 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-2ctxw" event={"ID":"db24d096-30dd-4099-82e8-fe87d671eb7d","Type":"ContainerStarted","Data":"a5d7a61478cb78c7e5945d9b03316707a3816210f9a80583a8b94bd095e7eabc"} Mar 14 09:27:52 crc kubenswrapper[4843]: I0314 09:27:52.073829 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-index-2ctxw" podStartSLOduration=1.873716382 podStartE2EDuration="2.073813501s" podCreationTimestamp="2026-03-14 09:27:50 +0000 UTC" firstStartedPulling="2026-03-14 09:27:51.442142743 +0000 UTC m=+1038.754753871" lastFinishedPulling="2026-03-14 09:27:51.642239862 +0000 UTC m=+1038.954850990" observedRunningTime="2026-03-14 09:27:52.070899641 +0000 UTC m=+1039.383510779" watchObservedRunningTime="2026-03-14 09:27:52.073813501 +0000 UTC m=+1039.386424619" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.161484 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-index-2ctxw"] Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.161690 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/watcher-operator-index-2ctxw" podUID="db24d096-30dd-4099-82e8-fe87d671eb7d" containerName="registry-server" containerID="cri-o://658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553" gracePeriod=2 Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.566740 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.732073 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hsd9\" (UniqueName: \"kubernetes.io/projected/db24d096-30dd-4099-82e8-fe87d671eb7d-kube-api-access-6hsd9\") pod \"db24d096-30dd-4099-82e8-fe87d671eb7d\" (UID: \"db24d096-30dd-4099-82e8-fe87d671eb7d\") " Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.738329 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db24d096-30dd-4099-82e8-fe87d671eb7d-kube-api-access-6hsd9" (OuterVolumeSpecName: "kube-api-access-6hsd9") pod "db24d096-30dd-4099-82e8-fe87d671eb7d" (UID: "db24d096-30dd-4099-82e8-fe87d671eb7d"). InnerVolumeSpecName "kube-api-access-6hsd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.777577 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-index-7zk59"] Mar 14 09:27:54 crc kubenswrapper[4843]: E0314 09:27:54.777920 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db24d096-30dd-4099-82e8-fe87d671eb7d" containerName="registry-server" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.777937 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="db24d096-30dd-4099-82e8-fe87d671eb7d" containerName="registry-server" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.778132 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="db24d096-30dd-4099-82e8-fe87d671eb7d" containerName="registry-server" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.778627 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.788357 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-index-7zk59"] Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.834004 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hsd9\" (UniqueName: \"kubernetes.io/projected/db24d096-30dd-4099-82e8-fe87d671eb7d-kube-api-access-6hsd9\") on node \"crc\" DevicePath \"\"" Mar 14 09:27:54 crc kubenswrapper[4843]: I0314 09:27:54.935142 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt8v9\" (UniqueName: \"kubernetes.io/projected/fa363db9-1707-497d-afab-dcf88242cdc2-kube-api-access-tt8v9\") pod \"watcher-operator-index-7zk59\" (UID: \"fa363db9-1707-497d-afab-dcf88242cdc2\") " pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.036818 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt8v9\" (UniqueName: \"kubernetes.io/projected/fa363db9-1707-497d-afab-dcf88242cdc2-kube-api-access-tt8v9\") pod \"watcher-operator-index-7zk59\" (UID: \"fa363db9-1707-497d-afab-dcf88242cdc2\") " pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.055666 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt8v9\" (UniqueName: \"kubernetes.io/projected/fa363db9-1707-497d-afab-dcf88242cdc2-kube-api-access-tt8v9\") pod \"watcher-operator-index-7zk59\" (UID: \"fa363db9-1707-497d-afab-dcf88242cdc2\") " pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.072754 4843 generic.go:334] "Generic (PLEG): container finished" podID="db24d096-30dd-4099-82e8-fe87d671eb7d" containerID="658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553" exitCode=0 Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.072816 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-2ctxw" event={"ID":"db24d096-30dd-4099-82e8-fe87d671eb7d","Type":"ContainerDied","Data":"658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553"} Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.072833 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-2ctxw" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.072847 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-2ctxw" event={"ID":"db24d096-30dd-4099-82e8-fe87d671eb7d","Type":"ContainerDied","Data":"a5d7a61478cb78c7e5945d9b03316707a3816210f9a80583a8b94bd095e7eabc"} Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.072867 4843 scope.go:117] "RemoveContainer" containerID="658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.100362 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.115942 4843 scope.go:117] "RemoveContainer" containerID="658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553" Mar 14 09:27:55 crc kubenswrapper[4843]: E0314 09:27:55.116367 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553\": container with ID starting with 658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553 not found: ID does not exist" containerID="658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.116407 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553"} err="failed to get container status \"658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553\": rpc error: code = NotFound desc = could not find container \"658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553\": container with ID starting with 658073bc19f2c33eeb9ec2d61b4f290ea7afabcaeeb59b66f6791c69fa78a553 not found: ID does not exist" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.121060 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-index-2ctxw"] Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.128850 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/watcher-operator-index-2ctxw"] Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.351690 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db24d096-30dd-4099-82e8-fe87d671eb7d" path="/var/lib/kubelet/pods/db24d096-30dd-4099-82e8-fe87d671eb7d/volumes" Mar 14 09:27:55 crc kubenswrapper[4843]: I0314 09:27:55.538653 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-index-7zk59"] Mar 14 09:27:55 crc kubenswrapper[4843]: W0314 09:27:55.541366 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa363db9_1707_497d_afab_dcf88242cdc2.slice/crio-723922a77d8f7e9150e4bd044c403baac44b4fbc5224abd63559aa42a2466f47 WatchSource:0}: Error finding container 723922a77d8f7e9150e4bd044c403baac44b4fbc5224abd63559aa42a2466f47: Status 404 returned error can't find the container with id 723922a77d8f7e9150e4bd044c403baac44b4fbc5224abd63559aa42a2466f47 Mar 14 09:27:56 crc kubenswrapper[4843]: I0314 09:27:56.083163 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-7zk59" event={"ID":"fa363db9-1707-497d-afab-dcf88242cdc2","Type":"ContainerStarted","Data":"53cd16dcebf298763ccf05f0130d8562f4fef3758514511dd5f0e58a9744b61b"} Mar 14 09:27:56 crc kubenswrapper[4843]: I0314 09:27:56.083230 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-index-7zk59" event={"ID":"fa363db9-1707-497d-afab-dcf88242cdc2","Type":"ContainerStarted","Data":"723922a77d8f7e9150e4bd044c403baac44b4fbc5224abd63559aa42a2466f47"} Mar 14 09:27:56 crc kubenswrapper[4843]: I0314 09:27:56.107467 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-index-7zk59" podStartSLOduration=2.056680287 podStartE2EDuration="2.107445064s" podCreationTimestamp="2026-03-14 09:27:54 +0000 UTC" firstStartedPulling="2026-03-14 09:27:55.545038647 +0000 UTC m=+1042.857649775" lastFinishedPulling="2026-03-14 09:27:55.595803424 +0000 UTC m=+1042.908414552" observedRunningTime="2026-03-14 09:27:56.105946348 +0000 UTC m=+1043.418557486" watchObservedRunningTime="2026-03-14 09:27:56.107445064 +0000 UTC m=+1043.420056192" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.134524 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558008-x58wg"] Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.136509 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.138579 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.139653 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.140728 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.143559 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558008-x58wg"] Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.211037 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brgbk\" (UniqueName: \"kubernetes.io/projected/6e5343d1-8bed-448c-a87f-a032b4f6a1a2-kube-api-access-brgbk\") pod \"auto-csr-approver-29558008-x58wg\" (UID: \"6e5343d1-8bed-448c-a87f-a032b4f6a1a2\") " pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.312310 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brgbk\" (UniqueName: \"kubernetes.io/projected/6e5343d1-8bed-448c-a87f-a032b4f6a1a2-kube-api-access-brgbk\") pod \"auto-csr-approver-29558008-x58wg\" (UID: \"6e5343d1-8bed-448c-a87f-a032b4f6a1a2\") " pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.346135 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brgbk\" (UniqueName: \"kubernetes.io/projected/6e5343d1-8bed-448c-a87f-a032b4f6a1a2-kube-api-access-brgbk\") pod \"auto-csr-approver-29558008-x58wg\" (UID: \"6e5343d1-8bed-448c-a87f-a032b4f6a1a2\") " pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.474738 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:00 crc kubenswrapper[4843]: I0314 09:28:00.924148 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558008-x58wg"] Mar 14 09:28:01 crc kubenswrapper[4843]: I0314 09:28:01.124022 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558008-x58wg" event={"ID":"6e5343d1-8bed-448c-a87f-a032b4f6a1a2","Type":"ContainerStarted","Data":"1954140de700d240857f45ba825bfa9530cc98bb5a428529efd3356d21ee302c"} Mar 14 09:28:02 crc kubenswrapper[4843]: I0314 09:28:02.131734 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558008-x58wg" event={"ID":"6e5343d1-8bed-448c-a87f-a032b4f6a1a2","Type":"ContainerStarted","Data":"84f73c8842bd583d796d0e6de35d870485969f1969e38951dc0297f05c21d115"} Mar 14 09:28:02 crc kubenswrapper[4843]: I0314 09:28:02.150650 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558008-x58wg" podStartSLOduration=1.300633284 podStartE2EDuration="2.150629748s" podCreationTimestamp="2026-03-14 09:28:00 +0000 UTC" firstStartedPulling="2026-03-14 09:28:00.93184232 +0000 UTC m=+1048.244453458" lastFinishedPulling="2026-03-14 09:28:01.781838794 +0000 UTC m=+1049.094449922" observedRunningTime="2026-03-14 09:28:02.144937101 +0000 UTC m=+1049.457548229" watchObservedRunningTime="2026-03-14 09:28:02.150629748 +0000 UTC m=+1049.463240876" Mar 14 09:28:03 crc kubenswrapper[4843]: I0314 09:28:03.143784 4843 generic.go:334] "Generic (PLEG): container finished" podID="6e5343d1-8bed-448c-a87f-a032b4f6a1a2" containerID="84f73c8842bd583d796d0e6de35d870485969f1969e38951dc0297f05c21d115" exitCode=0 Mar 14 09:28:03 crc kubenswrapper[4843]: I0314 09:28:03.143835 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558008-x58wg" event={"ID":"6e5343d1-8bed-448c-a87f-a032b4f6a1a2","Type":"ContainerDied","Data":"84f73c8842bd583d796d0e6de35d870485969f1969e38951dc0297f05c21d115"} Mar 14 09:28:04 crc kubenswrapper[4843]: I0314 09:28:04.454599 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:04 crc kubenswrapper[4843]: I0314 09:28:04.573382 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brgbk\" (UniqueName: \"kubernetes.io/projected/6e5343d1-8bed-448c-a87f-a032b4f6a1a2-kube-api-access-brgbk\") pod \"6e5343d1-8bed-448c-a87f-a032b4f6a1a2\" (UID: \"6e5343d1-8bed-448c-a87f-a032b4f6a1a2\") " Mar 14 09:28:04 crc kubenswrapper[4843]: I0314 09:28:04.580719 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5343d1-8bed-448c-a87f-a032b4f6a1a2-kube-api-access-brgbk" (OuterVolumeSpecName: "kube-api-access-brgbk") pod "6e5343d1-8bed-448c-a87f-a032b4f6a1a2" (UID: "6e5343d1-8bed-448c-a87f-a032b4f6a1a2"). InnerVolumeSpecName "kube-api-access-brgbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:28:04 crc kubenswrapper[4843]: I0314 09:28:04.674685 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brgbk\" (UniqueName: \"kubernetes.io/projected/6e5343d1-8bed-448c-a87f-a032b4f6a1a2-kube-api-access-brgbk\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.106543 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.108499 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.155668 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.161237 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558008-x58wg" event={"ID":"6e5343d1-8bed-448c-a87f-a032b4f6a1a2","Type":"ContainerDied","Data":"1954140de700d240857f45ba825bfa9530cc98bb5a428529efd3356d21ee302c"} Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.161287 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1954140de700d240857f45ba825bfa9530cc98bb5a428529efd3356d21ee302c" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.161251 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558008-x58wg" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.207036 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-index-7zk59" Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.214826 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558002-w844d"] Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.225562 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558002-w844d"] Mar 14 09:28:05 crc kubenswrapper[4843]: I0314 09:28:05.349200 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8" path="/var/lib/kubelet/pods/54d65ce7-9a3f-4fd6-8d9d-a1668ab57fd8/volumes" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.007593 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d"] Mar 14 09:28:07 crc kubenswrapper[4843]: E0314 09:28:07.008828 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5343d1-8bed-448c-a87f-a032b4f6a1a2" containerName="oc" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.008975 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5343d1-8bed-448c-a87f-a032b4f6a1a2" containerName="oc" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.009187 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5343d1-8bed-448c-a87f-a032b4f6a1a2" containerName="oc" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.010407 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.013424 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7m7z9" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.018065 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d"] Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.040464 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-util\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.040793 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-bundle\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.040936 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lksnn\" (UniqueName: \"kubernetes.io/projected/8cd388c8-6f9c-469d-b08f-32ad35850e12-kube-api-access-lksnn\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.142772 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-util\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.142845 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-bundle\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.142880 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lksnn\" (UniqueName: \"kubernetes.io/projected/8cd388c8-6f9c-469d-b08f-32ad35850e12-kube-api-access-lksnn\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.143646 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-util\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.143725 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-bundle\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.171208 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lksnn\" (UniqueName: \"kubernetes.io/projected/8cd388c8-6f9c-469d-b08f-32ad35850e12-kube-api-access-lksnn\") pod \"65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.327792 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:07 crc kubenswrapper[4843]: I0314 09:28:07.815696 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d"] Mar 14 09:28:07 crc kubenswrapper[4843]: W0314 09:28:07.820641 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cd388c8_6f9c_469d_b08f_32ad35850e12.slice/crio-1f1a2871b7375ad888619b0bf1f8b5f4901e381f34f58f69c6296e19a17cfc1f WatchSource:0}: Error finding container 1f1a2871b7375ad888619b0bf1f8b5f4901e381f34f58f69c6296e19a17cfc1f: Status 404 returned error can't find the container with id 1f1a2871b7375ad888619b0bf1f8b5f4901e381f34f58f69c6296e19a17cfc1f Mar 14 09:28:08 crc kubenswrapper[4843]: I0314 09:28:08.199070 4843 generic.go:334] "Generic (PLEG): container finished" podID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerID="725f05c022dc8905c65352565186562c4b7996205e290ca187eef1981afcbf00" exitCode=0 Mar 14 09:28:08 crc kubenswrapper[4843]: I0314 09:28:08.199261 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" event={"ID":"8cd388c8-6f9c-469d-b08f-32ad35850e12","Type":"ContainerDied","Data":"725f05c022dc8905c65352565186562c4b7996205e290ca187eef1981afcbf00"} Mar 14 09:28:08 crc kubenswrapper[4843]: I0314 09:28:08.200654 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" event={"ID":"8cd388c8-6f9c-469d-b08f-32ad35850e12","Type":"ContainerStarted","Data":"1f1a2871b7375ad888619b0bf1f8b5f4901e381f34f58f69c6296e19a17cfc1f"} Mar 14 09:28:09 crc kubenswrapper[4843]: I0314 09:28:09.210653 4843 generic.go:334] "Generic (PLEG): container finished" podID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerID="b2ebda10bf6718deee76c2f7a3ce2933109b325463a505c793f05d4f7f39a7ab" exitCode=0 Mar 14 09:28:09 crc kubenswrapper[4843]: I0314 09:28:09.210760 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" event={"ID":"8cd388c8-6f9c-469d-b08f-32ad35850e12","Type":"ContainerDied","Data":"b2ebda10bf6718deee76c2f7a3ce2933109b325463a505c793f05d4f7f39a7ab"} Mar 14 09:28:10 crc kubenswrapper[4843]: I0314 09:28:10.220653 4843 generic.go:334] "Generic (PLEG): container finished" podID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerID="4c6884ba3ddabbcdba1291e6a1fa44a6d7d841308863ac985fe0f11af52dc6d4" exitCode=0 Mar 14 09:28:10 crc kubenswrapper[4843]: I0314 09:28:10.220700 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" event={"ID":"8cd388c8-6f9c-469d-b08f-32ad35850e12","Type":"ContainerDied","Data":"4c6884ba3ddabbcdba1291e6a1fa44a6d7d841308863ac985fe0f11af52dc6d4"} Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.529558 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.609787 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lksnn\" (UniqueName: \"kubernetes.io/projected/8cd388c8-6f9c-469d-b08f-32ad35850e12-kube-api-access-lksnn\") pod \"8cd388c8-6f9c-469d-b08f-32ad35850e12\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.609857 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-util\") pod \"8cd388c8-6f9c-469d-b08f-32ad35850e12\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.609996 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-bundle\") pod \"8cd388c8-6f9c-469d-b08f-32ad35850e12\" (UID: \"8cd388c8-6f9c-469d-b08f-32ad35850e12\") " Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.611205 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-bundle" (OuterVolumeSpecName: "bundle") pod "8cd388c8-6f9c-469d-b08f-32ad35850e12" (UID: "8cd388c8-6f9c-469d-b08f-32ad35850e12"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.619490 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cd388c8-6f9c-469d-b08f-32ad35850e12-kube-api-access-lksnn" (OuterVolumeSpecName: "kube-api-access-lksnn") pod "8cd388c8-6f9c-469d-b08f-32ad35850e12" (UID: "8cd388c8-6f9c-469d-b08f-32ad35850e12"). InnerVolumeSpecName "kube-api-access-lksnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.623459 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-util" (OuterVolumeSpecName: "util") pod "8cd388c8-6f9c-469d-b08f-32ad35850e12" (UID: "8cd388c8-6f9c-469d-b08f-32ad35850e12"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.711307 4843 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.711631 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lksnn\" (UniqueName: \"kubernetes.io/projected/8cd388c8-6f9c-469d-b08f-32ad35850e12-kube-api-access-lksnn\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:11 crc kubenswrapper[4843]: I0314 09:28:11.711747 4843 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cd388c8-6f9c-469d-b08f-32ad35850e12-util\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:12 crc kubenswrapper[4843]: I0314 09:28:12.243961 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" event={"ID":"8cd388c8-6f9c-469d-b08f-32ad35850e12","Type":"ContainerDied","Data":"1f1a2871b7375ad888619b0bf1f8b5f4901e381f34f58f69c6296e19a17cfc1f"} Mar 14 09:28:12 crc kubenswrapper[4843]: I0314 09:28:12.243996 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d" Mar 14 09:28:12 crc kubenswrapper[4843]: I0314 09:28:12.244002 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f1a2871b7375ad888619b0bf1f8b5f4901e381f34f58f69c6296e19a17cfc1f" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.455713 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh"] Mar 14 09:28:19 crc kubenswrapper[4843]: E0314 09:28:19.457629 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="extract" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.457728 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="extract" Mar 14 09:28:19 crc kubenswrapper[4843]: E0314 09:28:19.457807 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="pull" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.457918 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="pull" Mar 14 09:28:19 crc kubenswrapper[4843]: E0314 09:28:19.458001 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="util" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.458068 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="util" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.458355 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd388c8-6f9c-469d-b08f-32ad35850e12" containerName="extract" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.459008 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.461354 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-vn982" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.462560 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-service-cert" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.470956 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh"] Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.627512 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-webhook-cert\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.627819 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-apiservice-cert\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.627893 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrh6s\" (UniqueName: \"kubernetes.io/projected/0f7fc64b-2d4b-467f-a932-632c30efeab5-kube-api-access-qrh6s\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.729610 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrh6s\" (UniqueName: \"kubernetes.io/projected/0f7fc64b-2d4b-467f-a932-632c30efeab5-kube-api-access-qrh6s\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.729694 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-webhook-cert\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.729721 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-apiservice-cert\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.735422 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-webhook-cert\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.738498 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-apiservice-cert\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.746757 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrh6s\" (UniqueName: \"kubernetes.io/projected/0f7fc64b-2d4b-467f-a932-632c30efeab5-kube-api-access-qrh6s\") pod \"watcher-operator-controller-manager-6d87bdbc68-f5xmh\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:19 crc kubenswrapper[4843]: I0314 09:28:19.780063 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:20 crc kubenswrapper[4843]: I0314 09:28:20.141965 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh"] Mar 14 09:28:20 crc kubenswrapper[4843]: I0314 09:28:20.308807 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" event={"ID":"0f7fc64b-2d4b-467f-a932-632c30efeab5","Type":"ContainerStarted","Data":"47c52503d0050eac4c6a40711d23a7f9f7f8c4b9a014c9be92c643c9fb68097b"} Mar 14 09:28:20 crc kubenswrapper[4843]: I0314 09:28:20.309006 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:20 crc kubenswrapper[4843]: I0314 09:28:20.331003 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" podStartSLOduration=1.330980019 podStartE2EDuration="1.330980019s" podCreationTimestamp="2026-03-14 09:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:28:20.329894663 +0000 UTC m=+1067.642505801" watchObservedRunningTime="2026-03-14 09:28:20.330980019 +0000 UTC m=+1067.643591147" Mar 14 09:28:21 crc kubenswrapper[4843]: I0314 09:28:21.319712 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" event={"ID":"0f7fc64b-2d4b-467f-a932-632c30efeab5","Type":"ContainerStarted","Data":"5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3"} Mar 14 09:28:29 crc kubenswrapper[4843]: I0314 09:28:29.784027 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:30 crc kubenswrapper[4843]: I0314 09:28:30.967500 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg"] Mar 14 09:28:30 crc kubenswrapper[4843]: I0314 09:28:30.969820 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:30 crc kubenswrapper[4843]: I0314 09:28:30.991249 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg"] Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.088007 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8b453c13-357e-4996-95ba-02e6d834a5b8-apiservice-cert\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.088071 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8b453c13-357e-4996-95ba-02e6d834a5b8-webhook-cert\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.088142 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64mkx\" (UniqueName: \"kubernetes.io/projected/8b453c13-357e-4996-95ba-02e6d834a5b8-kube-api-access-64mkx\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.188949 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64mkx\" (UniqueName: \"kubernetes.io/projected/8b453c13-357e-4996-95ba-02e6d834a5b8-kube-api-access-64mkx\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.189101 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8b453c13-357e-4996-95ba-02e6d834a5b8-apiservice-cert\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.189152 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8b453c13-357e-4996-95ba-02e6d834a5b8-webhook-cert\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.196668 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8b453c13-357e-4996-95ba-02e6d834a5b8-webhook-cert\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.196700 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8b453c13-357e-4996-95ba-02e6d834a5b8-apiservice-cert\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.210060 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64mkx\" (UniqueName: \"kubernetes.io/projected/8b453c13-357e-4996-95ba-02e6d834a5b8-kube-api-access-64mkx\") pod \"watcher-operator-controller-manager-66c89757b4-kgcsg\" (UID: \"8b453c13-357e-4996-95ba-02e6d834a5b8\") " pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.303112 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:31 crc kubenswrapper[4843]: I0314 09:28:31.774442 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg"] Mar 14 09:28:32 crc kubenswrapper[4843]: I0314 09:28:32.410845 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" event={"ID":"8b453c13-357e-4996-95ba-02e6d834a5b8","Type":"ContainerStarted","Data":"613f6ee35350453773333648bef345e2ceca7070368d53e9e703359dec38300b"} Mar 14 09:28:32 crc kubenswrapper[4843]: I0314 09:28:32.410892 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" event={"ID":"8b453c13-357e-4996-95ba-02e6d834a5b8","Type":"ContainerStarted","Data":"ff020fda744b148cbf54a5d4bd8615444509d6dce4a01f0338d3928b6ec02c22"} Mar 14 09:28:32 crc kubenswrapper[4843]: I0314 09:28:32.411175 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:32 crc kubenswrapper[4843]: I0314 09:28:32.443017 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" podStartSLOduration=2.442995904 podStartE2EDuration="2.442995904s" podCreationTimestamp="2026-03-14 09:28:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:28:32.435509144 +0000 UTC m=+1079.748120282" watchObservedRunningTime="2026-03-14 09:28:32.442995904 +0000 UTC m=+1079.755607032" Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.307739 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-66c89757b4-kgcsg" Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.382683 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh"] Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.382927 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" podUID="0f7fc64b-2d4b-467f-a932-632c30efeab5" containerName="manager" containerID="cri-o://5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3" gracePeriod=10 Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.809486 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.948609 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-webhook-cert\") pod \"0f7fc64b-2d4b-467f-a932-632c30efeab5\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.948662 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrh6s\" (UniqueName: \"kubernetes.io/projected/0f7fc64b-2d4b-467f-a932-632c30efeab5-kube-api-access-qrh6s\") pod \"0f7fc64b-2d4b-467f-a932-632c30efeab5\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.948774 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-apiservice-cert\") pod \"0f7fc64b-2d4b-467f-a932-632c30efeab5\" (UID: \"0f7fc64b-2d4b-467f-a932-632c30efeab5\") " Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.954389 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "0f7fc64b-2d4b-467f-a932-632c30efeab5" (UID: "0f7fc64b-2d4b-467f-a932-632c30efeab5"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.954472 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f7fc64b-2d4b-467f-a932-632c30efeab5-kube-api-access-qrh6s" (OuterVolumeSpecName: "kube-api-access-qrh6s") pod "0f7fc64b-2d4b-467f-a932-632c30efeab5" (UID: "0f7fc64b-2d4b-467f-a932-632c30efeab5"). InnerVolumeSpecName "kube-api-access-qrh6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:28:41 crc kubenswrapper[4843]: I0314 09:28:41.954928 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "0f7fc64b-2d4b-467f-a932-632c30efeab5" (UID: "0f7fc64b-2d4b-467f-a932-632c30efeab5"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.050226 4843 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.050290 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrh6s\" (UniqueName: \"kubernetes.io/projected/0f7fc64b-2d4b-467f-a932-632c30efeab5-kube-api-access-qrh6s\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.050305 4843 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f7fc64b-2d4b-467f-a932-632c30efeab5-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.481111 4843 generic.go:334] "Generic (PLEG): container finished" podID="0f7fc64b-2d4b-467f-a932-632c30efeab5" containerID="5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3" exitCode=0 Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.481178 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" event={"ID":"0f7fc64b-2d4b-467f-a932-632c30efeab5","Type":"ContainerDied","Data":"5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3"} Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.481229 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" event={"ID":"0f7fc64b-2d4b-467f-a932-632c30efeab5","Type":"ContainerDied","Data":"47c52503d0050eac4c6a40711d23a7f9f7f8c4b9a014c9be92c643c9fb68097b"} Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.481259 4843 scope.go:117] "RemoveContainer" containerID="5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.481182 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.520337 4843 scope.go:117] "RemoveContainer" containerID="5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3" Mar 14 09:28:42 crc kubenswrapper[4843]: E0314 09:28:42.520924 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3\": container with ID starting with 5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3 not found: ID does not exist" containerID="5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.520960 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3"} err="failed to get container status \"5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3\": rpc error: code = NotFound desc = could not find container \"5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3\": container with ID starting with 5e8ccd6e1b6063bf10901d2c775612cb4952bce486cfcbab7e1bf594653321a3 not found: ID does not exist" Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.523370 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh"] Mar 14 09:28:42 crc kubenswrapper[4843]: I0314 09:28:42.530108 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d87bdbc68-f5xmh"] Mar 14 09:28:43 crc kubenswrapper[4843]: I0314 09:28:43.350791 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f7fc64b-2d4b-467f-a932-632c30efeab5" path="/var/lib/kubelet/pods/0f7fc64b-2d4b-467f-a932-632c30efeab5/volumes" Mar 14 09:28:44 crc kubenswrapper[4843]: I0314 09:28:44.966671 4843 scope.go:117] "RemoveContainer" containerID="0686858794fd5e4868ec4687669da644f44286f475b98ae0fcbb07066f2f50de" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.700656 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/rabbitmq-notifications-server-0"] Mar 14 09:28:53 crc kubenswrapper[4843]: E0314 09:28:53.701582 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f7fc64b-2d4b-467f-a932-632c30efeab5" containerName="manager" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.701598 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f7fc64b-2d4b-467f-a932-632c30efeab5" containerName="manager" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.701775 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f7fc64b-2d4b-467f-a932-632c30efeab5" containerName="manager" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.702683 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.707931 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-rabbitmq-notifications-svc" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.708185 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-notifications-server-dockercfg-zf22v" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.708440 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-notifications-plugins-conf" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.709303 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openshift-service-ca.crt" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.710410 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-notifications-erlang-cookie" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.710475 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-notifications-default-user" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.711755 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-notifications-server-conf" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.716572 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"kube-root-ca.crt" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.717753 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-notifications-config-data" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.718386 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-notifications-server-0"] Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805314 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eaec015a-59c1-492c-9487-003e222f49fc-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805372 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805391 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79zrc\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-kube-api-access-79zrc\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805426 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eaec015a-59c1-492c-9487-003e222f49fc-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805508 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805534 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805553 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805578 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805595 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-881da295-24d8-487f-b039-0fedf6e04bd2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-881da295-24d8-487f-b039-0fedf6e04bd2\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805615 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.805635 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906492 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eaec015a-59c1-492c-9487-003e222f49fc-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906548 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906577 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906600 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906624 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906643 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-881da295-24d8-487f-b039-0fedf6e04bd2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-881da295-24d8-487f-b039-0fedf6e04bd2\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906664 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906685 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906713 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eaec015a-59c1-492c-9487-003e222f49fc-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906736 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.906755 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79zrc\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-kube-api-access-79zrc\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.907456 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.907840 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.908338 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.908819 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.909003 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eaec015a-59c1-492c-9487-003e222f49fc-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.912638 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.914124 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.916240 4843 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.916306 4843 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-881da295-24d8-487f-b039-0fedf6e04bd2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-881da295-24d8-487f-b039-0fedf6e04bd2\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/804653f34312a9751bf42237fd7eebf1ecad12661e2f1c689b5be77dcacd41ca/globalmount\"" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.918142 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eaec015a-59c1-492c-9487-003e222f49fc-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.930755 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/rabbitmq-server-0"] Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.932196 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.936323 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79zrc\" (UniqueName: \"kubernetes.io/projected/eaec015a-59c1-492c-9487-003e222f49fc-kube-api-access-79zrc\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.937663 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-erlang-cookie" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.937992 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-default-user" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.938179 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-plugins-conf" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.938304 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"rabbitmq-server-dockercfg-8mm6x" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.938364 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-server-conf" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.938498 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"rabbitmq-config-data" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.938617 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-rabbitmq-svc" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.939674 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eaec015a-59c1-492c-9487-003e222f49fc-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.956958 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-server-0"] Mar 14 09:28:53 crc kubenswrapper[4843]: I0314 09:28:53.964399 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-881da295-24d8-487f-b039-0fedf6e04bd2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-881da295-24d8-487f-b039-0fedf6e04bd2\") pod \"rabbitmq-notifications-server-0\" (UID: \"eaec015a-59c1-492c-9487-003e222f49fc\") " pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008224 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008324 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008439 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ced48f61-fae8-48aa-819f-c26f7296dac5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008497 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008645 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ced48f61-fae8-48aa-819f-c26f7296dac5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008725 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008747 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008768 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008797 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008815 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2s6p\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-kube-api-access-k2s6p\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.008872 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-config-data\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.019366 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110015 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-config-data\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110146 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110220 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110263 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ced48f61-fae8-48aa-819f-c26f7296dac5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110320 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110385 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ced48f61-fae8-48aa-819f-c26f7296dac5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110425 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110466 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110492 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110543 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.110563 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2s6p\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-kube-api-access-k2s6p\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.111292 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.111388 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-config-data\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.111733 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.111887 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ced48f61-fae8-48aa-819f-c26f7296dac5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.111997 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.116289 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.116868 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ced48f61-fae8-48aa-819f-c26f7296dac5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.122018 4843 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.122068 4843 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a37aab4877b17d59fdc032e8d342bf94771230f2a7222e6e1cb8bda46e8989a3/globalmount\"" pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.122778 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.126254 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ced48f61-fae8-48aa-819f-c26f7296dac5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.139836 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2s6p\" (UniqueName: \"kubernetes.io/projected/ced48f61-fae8-48aa-819f-c26f7296dac5-kube-api-access-k2s6p\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.171783 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32433811-9bbc-404b-9edf-cde59c3ec7e9\") pod \"rabbitmq-server-0\" (UID: \"ced48f61-fae8-48aa-819f-c26f7296dac5\") " pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.300721 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.560986 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-notifications-server-0"] Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.595583 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:28:54 crc kubenswrapper[4843]: I0314 09:28:54.740060 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/rabbitmq-server-0"] Mar 14 09:28:54 crc kubenswrapper[4843]: W0314 09:28:54.743197 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podced48f61_fae8_48aa_819f_c26f7296dac5.slice/crio-6509fcbf662a6feca04e7fd84c8738a5ac16a84a4c7930ede041802757d29839 WatchSource:0}: Error finding container 6509fcbf662a6feca04e7fd84c8738a5ac16a84a4c7930ede041802757d29839: Status 404 returned error can't find the container with id 6509fcbf662a6feca04e7fd84c8738a5ac16a84a4c7930ede041802757d29839 Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.269203 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/openstack-galera-0"] Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.271621 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.279493 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openstack-config-data" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.279695 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-galera-openstack-svc" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.279802 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openstack-scripts" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.279843 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstack-galera-0"] Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.280137 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"galera-openstack-dockercfg-r2stv" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.288734 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"combined-ca-bundle" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.333905 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.333984 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-config-data-default\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.334042 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-kolla-config\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.334122 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.334147 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.334168 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.334651 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.337216 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqdl6\" (UniqueName: \"kubernetes.io/projected/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-kube-api-access-nqdl6\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441208 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441324 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441395 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441443 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqdl6\" (UniqueName: \"kubernetes.io/projected/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-kube-api-access-nqdl6\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441541 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441570 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-config-data-default\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441618 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-kolla-config\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.441660 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.448494 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.448769 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.451074 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-config-data-default\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.451521 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-kolla-config\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.455519 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.456003 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.457713 4843 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.457753 4843 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a819537094f568ccd61890b7ef44bf107ae6d4fbfcc1ce5b30b75deb85a95bb2/globalmount\"" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.474014 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqdl6\" (UniqueName: \"kubernetes.io/projected/fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43-kube-api-access-nqdl6\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.515313 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6b77149-674d-40c1-91b1-cb4df59ea6b5\") pod \"openstack-galera-0\" (UID: \"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43\") " pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.523197 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.525106 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.534727 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"memcached-config-data" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.534908 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"memcached-memcached-dockercfg-85dpd" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.535020 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-memcached-svc" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.542262 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.608660 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.620392 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"ced48f61-fae8-48aa-819f-c26f7296dac5","Type":"ContainerStarted","Data":"6509fcbf662a6feca04e7fd84c8738a5ac16a84a4c7930ede041802757d29839"} Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.639520 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"eaec015a-59c1-492c-9487-003e222f49fc","Type":"ContainerStarted","Data":"06bfe29b6f8c3521df46d24c51a1a2f32fa9478a23d8802a4c5915f0fee5c468"} Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.644325 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.644396 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm8c9\" (UniqueName: \"kubernetes.io/projected/8c74421b-0ff7-450b-abcc-8e374ff891ac-kube-api-access-sm8c9\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.644461 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-kolla-config\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.644507 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-config-data\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.644549 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.746678 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-kolla-config\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.746736 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-config-data\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.746771 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.746805 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.746830 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm8c9\" (UniqueName: \"kubernetes.io/projected/8c74421b-0ff7-450b-abcc-8e374ff891ac-kube-api-access-sm8c9\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.747960 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-config-data\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.748144 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-kolla-config\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.759224 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.759523 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.768802 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm8c9\" (UniqueName: \"kubernetes.io/projected/8c74421b-0ff7-450b-abcc-8e374ff891ac-kube-api-access-sm8c9\") pod \"memcached-0\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.857120 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.857546 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.858314 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.861547 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"telemetry-ceilometer-dockercfg-sgtqb" Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.878797 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:28:55 crc kubenswrapper[4843]: I0314 09:28:55.953781 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whs92\" (UniqueName: \"kubernetes.io/projected/1c3ac2af-99e9-46e3-a447-5dcef51487ef-kube-api-access-whs92\") pod \"kube-state-metrics-0\" (UID: \"1c3ac2af-99e9-46e3-a447-5dcef51487ef\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.055770 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whs92\" (UniqueName: \"kubernetes.io/projected/1c3ac2af-99e9-46e3-a447-5dcef51487ef-kube-api-access-whs92\") pod \"kube-state-metrics-0\" (UID: \"1c3ac2af-99e9-46e3-a447-5dcef51487ef\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.077449 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whs92\" (UniqueName: \"kubernetes.io/projected/1c3ac2af-99e9-46e3-a447-5dcef51487ef-kube-api-access-whs92\") pod \"kube-state-metrics-0\" (UID: \"1c3ac2af-99e9-46e3-a447-5dcef51487ef\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.194136 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstack-galera-0"] Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.197219 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.505190 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.537006 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/alertmanager-metric-storage-0"] Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.538878 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.542422 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"metric-storage-alertmanager-dockercfg-vlpcr" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.542477 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-generated" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.542505 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-web-config" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.542655 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-cluster-tls-config" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.542655 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"alertmanager-metric-storage-tls-assets-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.560569 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/alertmanager-metric-storage-0"] Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576442 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/120dfd6f-cc76-42f9-83fb-954ae00d21c4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576473 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576528 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rq2d\" (UniqueName: \"kubernetes.io/projected/120dfd6f-cc76-42f9-83fb-954ae00d21c4-kube-api-access-2rq2d\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576554 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/120dfd6f-cc76-42f9-83fb-954ae00d21c4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576575 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576595 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.576612 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/120dfd6f-cc76-42f9-83fb-954ae00d21c4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.619896 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.660977 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"8c74421b-0ff7-450b-abcc-8e374ff891ac","Type":"ContainerStarted","Data":"7de2f648bb91a55acdcd842126af4a4779d6e22a129eefb89fe20979d2b405a2"} Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.663536 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"1c3ac2af-99e9-46e3-a447-5dcef51487ef","Type":"ContainerStarted","Data":"d458a056eebdc9a487d83ee2f986323f545797fdbbac141aafdd15369a9ba9a2"} Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.664856 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43","Type":"ContainerStarted","Data":"5bef61fbda9e071a7ee17a7d0587a3a648cd9079a79a0443b6be5c42b7012fcc"} Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.678915 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/120dfd6f-cc76-42f9-83fb-954ae00d21c4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.678958 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.679011 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rq2d\" (UniqueName: \"kubernetes.io/projected/120dfd6f-cc76-42f9-83fb-954ae00d21c4-kube-api-access-2rq2d\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.679041 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/120dfd6f-cc76-42f9-83fb-954ae00d21c4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.679064 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.679088 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.679105 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/120dfd6f-cc76-42f9-83fb-954ae00d21c4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.683230 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/120dfd6f-cc76-42f9-83fb-954ae00d21c4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.687922 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.688399 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.695347 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/120dfd6f-cc76-42f9-83fb-954ae00d21c4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.695555 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/120dfd6f-cc76-42f9-83fb-954ae00d21c4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.697006 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/120dfd6f-cc76-42f9-83fb-954ae00d21c4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.700146 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rq2d\" (UniqueName: \"kubernetes.io/projected/120dfd6f-cc76-42f9-83fb-954ae00d21c4-kube-api-access-2rq2d\") pod \"alertmanager-metric-storage-0\" (UID: \"120dfd6f-cc76-42f9-83fb-954ae00d21c4\") " pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.895898 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.958967 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7"] Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.959825 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.963791 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-rcj8t" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.964003 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Mar 14 09:28:56 crc kubenswrapper[4843]: I0314 09:28:56.978952 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7"] Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.094362 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgwp5\" (UniqueName: \"kubernetes.io/projected/2c318b20-fdd8-4ebb-817b-14f94d14580c-kube-api-access-pgwp5\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.094706 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c318b20-fdd8-4ebb-817b-14f94d14580c-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.172764 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.174850 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.181574 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-tls-assets-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.182062 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.182531 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"metric-storage-prometheus-dockercfg-d8lcx" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.182922 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.182936 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-web-config" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.182996 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.183094 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-2" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.183325 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-1" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.196340 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c318b20-fdd8-4ebb-817b-14f94d14580c-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.196555 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgwp5\" (UniqueName: \"kubernetes.io/projected/2c318b20-fdd8-4ebb-817b-14f94d14580c-kube-api-access-pgwp5\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: E0314 09:28:57.196785 4843 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Mar 14 09:28:57 crc kubenswrapper[4843]: E0314 09:28:57.196840 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c318b20-fdd8-4ebb-817b-14f94d14580c-serving-cert podName:2c318b20-fdd8-4ebb-817b-14f94d14580c nodeName:}" failed. No retries permitted until 2026-03-14 09:28:57.696822742 +0000 UTC m=+1105.009433870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/2c318b20-fdd8-4ebb-817b-14f94d14580c-serving-cert") pod "observability-ui-dashboards-66cbf594b5-k8bb7" (UID: "2c318b20-fdd8-4ebb-817b-14f94d14580c") : secret "observability-ui-dashboards" not found Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.209479 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.247091 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgwp5\" (UniqueName: \"kubernetes.io/projected/2c318b20-fdd8-4ebb-817b-14f94d14580c-kube-api-access-pgwp5\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299241 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fd5n\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-kube-api-access-4fd5n\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299401 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299427 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299468 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299524 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299549 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299579 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299593 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299630 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-config\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.299650 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cece7c09-8726-42db-9df7-5b79ed16ad84-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401614 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401695 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401736 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401760 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401779 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401832 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-config\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401848 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cece7c09-8726-42db-9df7-5b79ed16ad84-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401928 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fd5n\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-kube-api-access-4fd5n\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401946 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.401981 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.403792 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.407934 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.408368 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.428989 4843 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.429027 4843 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/08b800371242d647042cc38b2143ae04a3181ea2c70d960f9b416f4399b9e184/globalmount\"" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.445228 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cece7c09-8726-42db-9df7-5b79ed16ad84-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.446711 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.447747 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.453825 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.457048 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-config\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.465862 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fd5n\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-kube-api-access-4fd5n\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.529749 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7cff745649-s9lhf"] Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.530688 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.559610 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cff745649-s9lhf"] Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.606878 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-oauth-serving-cert\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.606931 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7khs6\" (UniqueName: \"kubernetes.io/projected/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-kube-api-access-7khs6\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.606970 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-oauth-config\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.606987 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-serving-cert\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.607010 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-trusted-ca-bundle\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.607037 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-config\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.607074 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-service-ca\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.621153 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709209 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-service-ca\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709333 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-oauth-serving-cert\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709370 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7khs6\" (UniqueName: \"kubernetes.io/projected/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-kube-api-access-7khs6\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709395 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c318b20-fdd8-4ebb-817b-14f94d14580c-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709418 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-oauth-config\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709434 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-serving-cert\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709459 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-trusted-ca-bundle\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.709482 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-config\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.710319 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-oauth-serving-cert\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.711198 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-config\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.711364 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-trusted-ca-bundle\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.711388 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-service-ca\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.717241 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c318b20-fdd8-4ebb-817b-14f94d14580c-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-k8bb7\" (UID: \"2c318b20-fdd8-4ebb-817b-14f94d14580c\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.718073 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-serving-cert\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.742126 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-console-oauth-config\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.754786 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7khs6\" (UniqueName: \"kubernetes.io/projected/c3ea7d60-60f0-415b-ad47-e0f383fb4fca-kube-api-access-7khs6\") pod \"console-7cff745649-s9lhf\" (UID: \"c3ea7d60-60f0-415b-ad47-e0f383fb4fca\") " pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.777231 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/alertmanager-metric-storage-0"] Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.802061 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.898705 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:28:57 crc kubenswrapper[4843]: I0314 09:28:57.916528 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" Mar 14 09:28:58 crc kubenswrapper[4843]: I0314 09:28:58.345794 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:28:58 crc kubenswrapper[4843]: W0314 09:28:58.485188 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcece7c09_8726_42db_9df7_5b79ed16ad84.slice/crio-b86b0b6a192246d8025ac0e359f619da702b8d897e4d3ccb6f0097988c7b2f9d WatchSource:0}: Error finding container b86b0b6a192246d8025ac0e359f619da702b8d897e4d3ccb6f0097988c7b2f9d: Status 404 returned error can't find the container with id b86b0b6a192246d8025ac0e359f619da702b8d897e4d3ccb6f0097988c7b2f9d Mar 14 09:28:58 crc kubenswrapper[4843]: I0314 09:28:58.685577 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerStarted","Data":"b86b0b6a192246d8025ac0e359f619da702b8d897e4d3ccb6f0097988c7b2f9d"} Mar 14 09:28:58 crc kubenswrapper[4843]: I0314 09:28:58.687401 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"120dfd6f-cc76-42f9-83fb-954ae00d21c4","Type":"ContainerStarted","Data":"e55a3b44b5a788c73204c89f790869e6e2471d8f21a6651e919b7f266cd6c3b8"} Mar 14 09:28:59 crc kubenswrapper[4843]: I0314 09:28:59.548142 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cff745649-s9lhf"] Mar 14 09:28:59 crc kubenswrapper[4843]: W0314 09:28:59.607372 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3ea7d60_60f0_415b_ad47_e0f383fb4fca.slice/crio-06ec6f18135369fa8962f2920768705f62d868a12e277e6fa324ede11d0d49a4 WatchSource:0}: Error finding container 06ec6f18135369fa8962f2920768705f62d868a12e277e6fa324ede11d0d49a4: Status 404 returned error can't find the container with id 06ec6f18135369fa8962f2920768705f62d868a12e277e6fa324ede11d0d49a4 Mar 14 09:28:59 crc kubenswrapper[4843]: I0314 09:28:59.617961 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7"] Mar 14 09:28:59 crc kubenswrapper[4843]: I0314 09:28:59.700022 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cff745649-s9lhf" event={"ID":"c3ea7d60-60f0-415b-ad47-e0f383fb4fca","Type":"ContainerStarted","Data":"06ec6f18135369fa8962f2920768705f62d868a12e277e6fa324ede11d0d49a4"} Mar 14 09:29:00 crc kubenswrapper[4843]: W0314 09:28:59.999875 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c318b20_fdd8_4ebb_817b_14f94d14580c.slice/crio-86ad5b220b00c776202e229273bfefa884e0571cba576312eee9bb76722dfe81 WatchSource:0}: Error finding container 86ad5b220b00c776202e229273bfefa884e0571cba576312eee9bb76722dfe81: Status 404 returned error can't find the container with id 86ad5b220b00c776202e229273bfefa884e0571cba576312eee9bb76722dfe81 Mar 14 09:29:00 crc kubenswrapper[4843]: I0314 09:29:00.721513 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" event={"ID":"2c318b20-fdd8-4ebb-817b-14f94d14580c","Type":"ContainerStarted","Data":"86ad5b220b00c776202e229273bfefa884e0571cba576312eee9bb76722dfe81"} Mar 14 09:29:10 crc kubenswrapper[4843]: E0314 09:29:10.817101 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Mar 14 09:29:10 crc kubenswrapper[4843]: E0314 09:29:10.817887 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5d8h59hddh669h545h544h7dh79hddh59ch567h588h5c7h85h94hcch5cch596h589hb9hb8h65ch68fh79h58ch5bfh597h59h64fhffhf9hbq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sm8c9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_watcher-kuttl-default(8c74421b-0ff7-450b-abcc-8e374ff891ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:29:10 crc kubenswrapper[4843]: E0314 09:29:10.819019 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/memcached-0" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" Mar 14 09:29:11 crc kubenswrapper[4843]: E0314 09:29:11.804558 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="watcher-kuttl-default/memcached-0" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" Mar 14 09:29:13 crc kubenswrapper[4843]: E0314 09:29:13.055182 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Mar 14 09:29:13 crc kubenswrapper[4843]: E0314 09:29:13.055408 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqdl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_watcher-kuttl-default(fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:29:13 crc kubenswrapper[4843]: E0314 09:29:13.056771 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/openstack-galera-0" podUID="fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43" Mar 14 09:29:13 crc kubenswrapper[4843]: I0314 09:29:13.833885 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:29:13 crc kubenswrapper[4843]: I0314 09:29:13.838884 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" event={"ID":"2c318b20-fdd8-4ebb-817b-14f94d14580c","Type":"ContainerStarted","Data":"215c25050fa9f0612badf44a95362a7b15b3f97267b6c3d9dcc2bfa669dc133e"} Mar 14 09:29:13 crc kubenswrapper[4843]: I0314 09:29:13.841183 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cff745649-s9lhf" event={"ID":"c3ea7d60-60f0-415b-ad47-e0f383fb4fca","Type":"ContainerStarted","Data":"e37a9fd8f8fb047187d9ab3c7725f85d18bc013a7b43d858e7067ebd50739d3a"} Mar 14 09:29:13 crc kubenswrapper[4843]: E0314 09:29:13.842919 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="watcher-kuttl-default/openstack-galera-0" podUID="fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43" Mar 14 09:29:13 crc kubenswrapper[4843]: I0314 09:29:13.851054 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/kube-state-metrics-0" podStartSLOduration=2.4124204430000002 podStartE2EDuration="18.851035745s" podCreationTimestamp="2026-03-14 09:28:55 +0000 UTC" firstStartedPulling="2026-03-14 09:28:56.625885313 +0000 UTC m=+1103.938496441" lastFinishedPulling="2026-03-14 09:29:13.064500605 +0000 UTC m=+1120.377111743" observedRunningTime="2026-03-14 09:29:13.849169109 +0000 UTC m=+1121.161780247" watchObservedRunningTime="2026-03-14 09:29:13.851035745 +0000 UTC m=+1121.163646873" Mar 14 09:29:13 crc kubenswrapper[4843]: I0314 09:29:13.903225 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7cff745649-s9lhf" podStartSLOduration=16.90320175 podStartE2EDuration="16.90320175s" podCreationTimestamp="2026-03-14 09:28:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:29:13.89879704 +0000 UTC m=+1121.211408178" watchObservedRunningTime="2026-03-14 09:29:13.90320175 +0000 UTC m=+1121.215812878" Mar 14 09:29:13 crc kubenswrapper[4843]: I0314 09:29:13.923159 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-k8bb7" podStartSLOduration=4.880473286 podStartE2EDuration="17.923136368s" podCreationTimestamp="2026-03-14 09:28:56 +0000 UTC" firstStartedPulling="2026-03-14 09:29:00.002334205 +0000 UTC m=+1107.314945333" lastFinishedPulling="2026-03-14 09:29:13.044997267 +0000 UTC m=+1120.357608415" observedRunningTime="2026-03-14 09:29:13.912738198 +0000 UTC m=+1121.225349326" watchObservedRunningTime="2026-03-14 09:29:13.923136368 +0000 UTC m=+1121.235747496" Mar 14 09:29:14 crc kubenswrapper[4843]: I0314 09:29:14.849697 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"1c3ac2af-99e9-46e3-a447-5dcef51487ef","Type":"ContainerStarted","Data":"1d8d4d630e57b45c6c556e6b8945aa8ec049c786c5df18dac0aa554416c5882d"} Mar 14 09:29:15 crc kubenswrapper[4843]: I0314 09:29:15.387368 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:29:15 crc kubenswrapper[4843]: I0314 09:29:15.387670 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:29:15 crc kubenswrapper[4843]: I0314 09:29:15.859892 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"eaec015a-59c1-492c-9487-003e222f49fc","Type":"ContainerStarted","Data":"5463bfffb3cde618dd578fbf4fd21c1fede480401b70e9371b240d8444e028ed"} Mar 14 09:29:15 crc kubenswrapper[4843]: I0314 09:29:15.861143 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"ced48f61-fae8-48aa-819f-c26f7296dac5","Type":"ContainerStarted","Data":"780596fd9834bc6ac086d6ae972e3925650df9e8a64ebd8b32acd1091ce1aac8"} Mar 14 09:29:16 crc kubenswrapper[4843]: I0314 09:29:16.869089 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerStarted","Data":"ece2e21a80fd7340f1025c87ef67efe1eaa5da14b20b7b7dc7a231460fe7b095"} Mar 14 09:29:16 crc kubenswrapper[4843]: I0314 09:29:16.870387 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"120dfd6f-cc76-42f9-83fb-954ae00d21c4","Type":"ContainerStarted","Data":"89656880ea5e30f04618c3d64d4560f289108b9c5f091b8bbfd54bdb121db82f"} Mar 14 09:29:17 crc kubenswrapper[4843]: I0314 09:29:17.899782 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:29:17 crc kubenswrapper[4843]: I0314 09:29:17.899841 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:29:17 crc kubenswrapper[4843]: I0314 09:29:17.906060 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:29:18 crc kubenswrapper[4843]: I0314 09:29:18.888899 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7cff745649-s9lhf" Mar 14 09:29:19 crc kubenswrapper[4843]: I0314 09:29:19.003161 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-584b867db4-62h25"] Mar 14 09:29:22 crc kubenswrapper[4843]: I0314 09:29:22.915476 4843 generic.go:334] "Generic (PLEG): container finished" podID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerID="ece2e21a80fd7340f1025c87ef67efe1eaa5da14b20b7b7dc7a231460fe7b095" exitCode=0 Mar 14 09:29:22 crc kubenswrapper[4843]: I0314 09:29:22.915574 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerDied","Data":"ece2e21a80fd7340f1025c87ef67efe1eaa5da14b20b7b7dc7a231460fe7b095"} Mar 14 09:29:22 crc kubenswrapper[4843]: I0314 09:29:22.918541 4843 generic.go:334] "Generic (PLEG): container finished" podID="120dfd6f-cc76-42f9-83fb-954ae00d21c4" containerID="89656880ea5e30f04618c3d64d4560f289108b9c5f091b8bbfd54bdb121db82f" exitCode=0 Mar 14 09:29:22 crc kubenswrapper[4843]: I0314 09:29:22.918579 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"120dfd6f-cc76-42f9-83fb-954ae00d21c4","Type":"ContainerDied","Data":"89656880ea5e30f04618c3d64d4560f289108b9c5f091b8bbfd54bdb121db82f"} Mar 14 09:29:25 crc kubenswrapper[4843]: I0314 09:29:25.941563 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"8c74421b-0ff7-450b-abcc-8e374ff891ac","Type":"ContainerStarted","Data":"fc89194c3dae86995ed493f4a9bcad9fe005794878d0df571dbc3e997178bcb8"} Mar 14 09:29:25 crc kubenswrapper[4843]: I0314 09:29:25.942032 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/memcached-0" Mar 14 09:29:25 crc kubenswrapper[4843]: I0314 09:29:25.944594 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"120dfd6f-cc76-42f9-83fb-954ae00d21c4","Type":"ContainerStarted","Data":"f1add96750186830d9fbb363c0bf1c14c93f2c0219680d6f0a786403fd8ae284"} Mar 14 09:29:25 crc kubenswrapper[4843]: I0314 09:29:25.971510 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/memcached-0" podStartSLOduration=2.286194428 podStartE2EDuration="30.971485535s" podCreationTimestamp="2026-03-14 09:28:55 +0000 UTC" firstStartedPulling="2026-03-14 09:28:56.530455527 +0000 UTC m=+1103.843066655" lastFinishedPulling="2026-03-14 09:29:25.215746634 +0000 UTC m=+1132.528357762" observedRunningTime="2026-03-14 09:29:25.961203457 +0000 UTC m=+1133.273814595" watchObservedRunningTime="2026-03-14 09:29:25.971485535 +0000 UTC m=+1133.284096683" Mar 14 09:29:26 crc kubenswrapper[4843]: I0314 09:29:26.202841 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:29:27 crc kubenswrapper[4843]: I0314 09:29:27.966783 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/alertmanager-metric-storage-0" event={"ID":"120dfd6f-cc76-42f9-83fb-954ae00d21c4","Type":"ContainerStarted","Data":"8bbb250ca8ab83d5a9fcc79d823a356db859c10c72e94a1649e66fdf2686fa88"} Mar 14 09:29:27 crc kubenswrapper[4843]: I0314 09:29:27.967182 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:29:28 crc kubenswrapper[4843]: I0314 09:29:28.001254 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/alertmanager-metric-storage-0" podStartSLOduration=4.575211282 podStartE2EDuration="32.001234076s" podCreationTimestamp="2026-03-14 09:28:56 +0000 UTC" firstStartedPulling="2026-03-14 09:28:57.795604817 +0000 UTC m=+1105.108215945" lastFinishedPulling="2026-03-14 09:29:25.221627611 +0000 UTC m=+1132.534238739" observedRunningTime="2026-03-14 09:29:27.998098428 +0000 UTC m=+1135.310709556" watchObservedRunningTime="2026-03-14 09:29:28.001234076 +0000 UTC m=+1135.313845204" Mar 14 09:29:28 crc kubenswrapper[4843]: I0314 09:29:28.976337 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/alertmanager-metric-storage-0" Mar 14 09:29:30 crc kubenswrapper[4843]: I0314 09:29:30.859427 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/memcached-0" Mar 14 09:29:30 crc kubenswrapper[4843]: I0314 09:29:30.990686 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43","Type":"ContainerStarted","Data":"5ffe0e7550603ad23a8815498a7c09b680e686b82f3f1269c01a2aaa7e607f3b"} Mar 14 09:29:30 crc kubenswrapper[4843]: I0314 09:29:30.996419 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerStarted","Data":"3ce276aa60ba1d9105597c8d76ea3fc2407dc4c16576f315765a6db7916cf1b2"} Mar 14 09:29:33 crc kubenswrapper[4843]: I0314 09:29:33.019927 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerStarted","Data":"b90bdb90d3afdd048a8d5ef9eb97337323b5a17802c8516d1c9264699846c07c"} Mar 14 09:29:34 crc kubenswrapper[4843]: I0314 09:29:34.028374 4843 generic.go:334] "Generic (PLEG): container finished" podID="fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43" containerID="5ffe0e7550603ad23a8815498a7c09b680e686b82f3f1269c01a2aaa7e607f3b" exitCode=0 Mar 14 09:29:34 crc kubenswrapper[4843]: I0314 09:29:34.028438 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43","Type":"ContainerDied","Data":"5ffe0e7550603ad23a8815498a7c09b680e686b82f3f1269c01a2aaa7e607f3b"} Mar 14 09:29:36 crc kubenswrapper[4843]: I0314 09:29:36.061066 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstack-galera-0" event={"ID":"fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43","Type":"ContainerStarted","Data":"bb7e56bbefda47c375a3eeb4e8a3df6af87c6b9a2349983c16c8c6273c98a61d"} Mar 14 09:29:36 crc kubenswrapper[4843]: I0314 09:29:36.087152 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/openstack-galera-0" podStartSLOduration=8.301642408 podStartE2EDuration="42.087130046s" podCreationTimestamp="2026-03-14 09:28:54 +0000 UTC" firstStartedPulling="2026-03-14 09:28:56.251325986 +0000 UTC m=+1103.563937114" lastFinishedPulling="2026-03-14 09:29:30.036813624 +0000 UTC m=+1137.349424752" observedRunningTime="2026-03-14 09:29:36.080371316 +0000 UTC m=+1143.392982454" watchObservedRunningTime="2026-03-14 09:29:36.087130046 +0000 UTC m=+1143.399741174" Mar 14 09:29:38 crc kubenswrapper[4843]: I0314 09:29:38.079989 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerStarted","Data":"459a105316339e3d5ebac6d3b5d2d8d6062fc4e719a1d86db112f21914abfa97"} Mar 14 09:29:38 crc kubenswrapper[4843]: I0314 09:29:38.108672 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/prometheus-metric-storage-0" podStartSLOduration=3.320905862 podStartE2EDuration="42.108653792s" podCreationTimestamp="2026-03-14 09:28:56 +0000 UTC" firstStartedPulling="2026-03-14 09:28:58.498773682 +0000 UTC m=+1105.811384810" lastFinishedPulling="2026-03-14 09:29:37.286521612 +0000 UTC m=+1144.599132740" observedRunningTime="2026-03-14 09:29:38.105678428 +0000 UTC m=+1145.418289556" watchObservedRunningTime="2026-03-14 09:29:38.108653792 +0000 UTC m=+1145.421264920" Mar 14 09:29:42 crc kubenswrapper[4843]: I0314 09:29:42.802980 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:42 crc kubenswrapper[4843]: I0314 09:29:42.803590 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:42 crc kubenswrapper[4843]: I0314 09:29:42.805820 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:43 crc kubenswrapper[4843]: I0314 09:29:43.114663 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.047239 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-584b867db4-62h25" podUID="ab35bbcf-2757-4b2e-8a4d-7a716669c229" containerName="console" containerID="cri-o://7759b174c9c43bdc8815200b690a32544cb33e2859bb7f6e40d327e1baad80a4" gracePeriod=15 Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.584756 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-584b867db4-62h25_ab35bbcf-2757-4b2e-8a4d-7a716669c229/console/0.log" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.585098 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726533 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-trusted-ca-bundle\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726641 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nghsq\" (UniqueName: \"kubernetes.io/projected/ab35bbcf-2757-4b2e-8a4d-7a716669c229-kube-api-access-nghsq\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726772 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-service-ca\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726790 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-config\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726842 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-serving-cert\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726860 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-oauth-config\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.726887 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-oauth-serving-cert\") pod \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\" (UID: \"ab35bbcf-2757-4b2e-8a4d-7a716669c229\") " Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.727394 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.727413 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-service-ca" (OuterVolumeSpecName: "service-ca") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.727458 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-config" (OuterVolumeSpecName: "console-config") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.727467 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.731987 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.732412 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.732676 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab35bbcf-2757-4b2e-8a4d-7a716669c229-kube-api-access-nghsq" (OuterVolumeSpecName: "kube-api-access-nghsq") pod "ab35bbcf-2757-4b2e-8a4d-7a716669c229" (UID: "ab35bbcf-2757-4b2e-8a4d-7a716669c229"). InnerVolumeSpecName "kube-api-access-nghsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828593 4843 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828639 4843 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828653 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nghsq\" (UniqueName: \"kubernetes.io/projected/ab35bbcf-2757-4b2e-8a4d-7a716669c229-kube-api-access-nghsq\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828667 4843 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-service-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828678 4843 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828690 4843 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:44 crc kubenswrapper[4843]: I0314 09:29:44.828702 4843 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab35bbcf-2757-4b2e-8a4d-7a716669c229-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.060916 4843 scope.go:117] "RemoveContainer" containerID="7759b174c9c43bdc8815200b690a32544cb33e2859bb7f6e40d327e1baad80a4" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.142134 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-584b867db4-62h25" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.142133 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-584b867db4-62h25" event={"ID":"ab35bbcf-2757-4b2e-8a4d-7a716669c229","Type":"ContainerDied","Data":"7759b174c9c43bdc8815200b690a32544cb33e2859bb7f6e40d327e1baad80a4"} Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.142299 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-584b867db4-62h25" event={"ID":"ab35bbcf-2757-4b2e-8a4d-7a716669c229","Type":"ContainerDied","Data":"f8b38cb17dd8d4896cb805c622dca520de6f37f7fd8a1aca944cb15ead624fa9"} Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.178484 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-584b867db4-62h25"] Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.187603 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-584b867db4-62h25"] Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.312413 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.312746 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/prometheus-metric-storage-0" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="prometheus" containerID="cri-o://3ce276aa60ba1d9105597c8d76ea3fc2407dc4c16576f315765a6db7916cf1b2" gracePeriod=600 Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.312844 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/prometheus-metric-storage-0" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="thanos-sidecar" containerID="cri-o://459a105316339e3d5ebac6d3b5d2d8d6062fc4e719a1d86db112f21914abfa97" gracePeriod=600 Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.312893 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/prometheus-metric-storage-0" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="config-reloader" containerID="cri-o://b90bdb90d3afdd048a8d5ef9eb97337323b5a17802c8516d1c9264699846c07c" gracePeriod=600 Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.348477 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab35bbcf-2757-4b2e-8a4d-7a716669c229" path="/var/lib/kubelet/pods/ab35bbcf-2757-4b2e-8a4d-7a716669c229/volumes" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.387555 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.387618 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.610057 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.610145 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:29:45 crc kubenswrapper[4843]: I0314 09:29:45.706727 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.156741 4843 generic.go:334] "Generic (PLEG): container finished" podID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerID="459a105316339e3d5ebac6d3b5d2d8d6062fc4e719a1d86db112f21914abfa97" exitCode=0 Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.156775 4843 generic.go:334] "Generic (PLEG): container finished" podID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerID="b90bdb90d3afdd048a8d5ef9eb97337323b5a17802c8516d1c9264699846c07c" exitCode=0 Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.156785 4843 generic.go:334] "Generic (PLEG): container finished" podID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerID="3ce276aa60ba1d9105597c8d76ea3fc2407dc4c16576f315765a6db7916cf1b2" exitCode=0 Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.156826 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerDied","Data":"459a105316339e3d5ebac6d3b5d2d8d6062fc4e719a1d86db112f21914abfa97"} Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.156870 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerDied","Data":"b90bdb90d3afdd048a8d5ef9eb97337323b5a17802c8516d1c9264699846c07c"} Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.156884 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerDied","Data":"3ce276aa60ba1d9105597c8d76ea3fc2407dc4c16576f315765a6db7916cf1b2"} Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.240161 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.311585 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/openstack-galera-0" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.350215 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-config\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.350263 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-web-config\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.350312 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-tls-assets\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.350334 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cece7c09-8726-42db-9df7-5b79ed16ad84-config-out\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.351099 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.351140 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-thanos-prometheus-http-client-file\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.351215 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-2\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.351288 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fd5n\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-kube-api-access-4fd5n\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.351329 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-1\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.351356 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-0\") pod \"cece7c09-8726-42db-9df7-5b79ed16ad84\" (UID: \"cece7c09-8726-42db-9df7-5b79ed16ad84\") " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.352298 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.353689 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.353213 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.356718 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-kube-api-access-4fd5n" (OuterVolumeSpecName: "kube-api-access-4fd5n") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "kube-api-access-4fd5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.357780 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-config" (OuterVolumeSpecName: "config") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.371114 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.371172 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.382671 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cece7c09-8726-42db-9df7-5b79ed16ad84-config-out" (OuterVolumeSpecName: "config-out") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.407659 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-web-config" (OuterVolumeSpecName: "web-config") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.412804 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "cece7c09-8726-42db-9df7-5b79ed16ad84" (UID: "cece7c09-8726-42db-9df7-5b79ed16ad84"). InnerVolumeSpecName "pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453447 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fd5n\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-kube-api-access-4fd5n\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453485 4843 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453585 4843 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453600 4843 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453613 4843 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-web-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453622 4843 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cece7c09-8726-42db-9df7-5b79ed16ad84-tls-assets\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453631 4843 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cece7c09-8726-42db-9df7-5b79ed16ad84-config-out\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453658 4843 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") on node \"crc\" " Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453670 4843 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/cece7c09-8726-42db-9df7-5b79ed16ad84-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.453684 4843 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/cece7c09-8726-42db-9df7-5b79ed16ad84-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.472773 4843 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.472930 4843 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686") on node "crc" Mar 14 09:29:46 crc kubenswrapper[4843]: I0314 09:29:46.555512 4843 reconciler_common.go:293] "Volume detached for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.166064 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"cece7c09-8726-42db-9df7-5b79ed16ad84","Type":"ContainerDied","Data":"b86b0b6a192246d8025ac0e359f619da702b8d897e4d3ccb6f0097988c7b2f9d"} Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.166127 4843 scope.go:117] "RemoveContainer" containerID="459a105316339e3d5ebac6d3b5d2d8d6062fc4e719a1d86db112f21914abfa97" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.166148 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.200515 4843 scope.go:117] "RemoveContainer" containerID="b90bdb90d3afdd048a8d5ef9eb97337323b5a17802c8516d1c9264699846c07c" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.218448 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.220812 4843 scope.go:117] "RemoveContainer" containerID="3ce276aa60ba1d9105597c8d76ea3fc2407dc4c16576f315765a6db7916cf1b2" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.229994 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.241419 4843 scope.go:117] "RemoveContainer" containerID="ece2e21a80fd7340f1025c87ef67efe1eaa5da14b20b7b7dc7a231460fe7b095" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.268563 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:29:47 crc kubenswrapper[4843]: E0314 09:29:47.268881 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="init-config-reloader" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.268898 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="init-config-reloader" Mar 14 09:29:47 crc kubenswrapper[4843]: E0314 09:29:47.268916 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="thanos-sidecar" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.268925 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="thanos-sidecar" Mar 14 09:29:47 crc kubenswrapper[4843]: E0314 09:29:47.268937 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="config-reloader" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.268944 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="config-reloader" Mar 14 09:29:47 crc kubenswrapper[4843]: E0314 09:29:47.268951 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab35bbcf-2757-4b2e-8a4d-7a716669c229" containerName="console" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.268959 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab35bbcf-2757-4b2e-8a4d-7a716669c229" containerName="console" Mar 14 09:29:47 crc kubenswrapper[4843]: E0314 09:29:47.268996 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="prometheus" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.269004 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="prometheus" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.269168 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="thanos-sidecar" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.269183 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab35bbcf-2757-4b2e-8a4d-7a716669c229" containerName="console" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.269190 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="config-reloader" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.269202 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" containerName="prometheus" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.270499 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276114 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276389 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-metric-storage-prometheus-svc" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276399 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"metric-storage-prometheus-dockercfg-d8lcx" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276581 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-2" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276633 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276692 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.276727 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-web-config" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.281617 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"prometheus-metric-storage-rulefiles-1" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.313054 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"prometheus-metric-storage-tls-assets-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.367931 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368022 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368066 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368095 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368128 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgr9\" (UniqueName: \"kubernetes.io/projected/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-kube-api-access-vqgr9\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368156 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368187 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.368234 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.369157 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.380082 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-config\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.380200 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.380241 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.380386 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.441075 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cece7c09-8726-42db-9df7-5b79ed16ad84" path="/var/lib/kubelet/pods/cece7c09-8726-42db-9df7-5b79ed16ad84/volumes" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.441854 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484171 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-config\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484229 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484255 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484326 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484357 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484431 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484470 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484486 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484507 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgr9\" (UniqueName: \"kubernetes.io/projected/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-kube-api-access-vqgr9\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484541 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484561 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484614 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.484645 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.485605 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.492837 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.499418 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.516909 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.517433 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.518041 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.518281 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.525599 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.526583 4843 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.526692 4843 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/08b800371242d647042cc38b2143ae04a3181ea2c70d960f9b416f4399b9e184/globalmount\"" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.528656 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.528819 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgr9\" (UniqueName: \"kubernetes.io/projected/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-kube-api-access-vqgr9\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.539240 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-config\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.539261 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.576405 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faffcef0-1b9c-4b7e-8ba9-9224c9f55686\") pod \"prometheus-metric-storage-0\" (UID: \"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2\") " pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.696712 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:29:47 crc kubenswrapper[4843]: I0314 09:29:47.921114 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/prometheus-metric-storage-0"] Mar 14 09:29:48 crc kubenswrapper[4843]: I0314 09:29:48.178806 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2","Type":"ContainerStarted","Data":"0d428f2e0d5151c805d2c92a17ce2abc70d48b7a0c4620172c53ee5fda06bbba"} Mar 14 09:29:48 crc kubenswrapper[4843]: I0314 09:29:48.181983 4843 generic.go:334] "Generic (PLEG): container finished" podID="ced48f61-fae8-48aa-819f-c26f7296dac5" containerID="780596fd9834bc6ac086d6ae972e3925650df9e8a64ebd8b32acd1091ce1aac8" exitCode=0 Mar 14 09:29:48 crc kubenswrapper[4843]: I0314 09:29:48.182046 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"ced48f61-fae8-48aa-819f-c26f7296dac5","Type":"ContainerDied","Data":"780596fd9834bc6ac086d6ae972e3925650df9e8a64ebd8b32acd1091ce1aac8"} Mar 14 09:29:48 crc kubenswrapper[4843]: I0314 09:29:48.186503 4843 generic.go:334] "Generic (PLEG): container finished" podID="eaec015a-59c1-492c-9487-003e222f49fc" containerID="5463bfffb3cde618dd578fbf4fd21c1fede480401b70e9371b240d8444e028ed" exitCode=0 Mar 14 09:29:48 crc kubenswrapper[4843]: I0314 09:29:48.187333 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"eaec015a-59c1-492c-9487-003e222f49fc","Type":"ContainerDied","Data":"5463bfffb3cde618dd578fbf4fd21c1fede480401b70e9371b240d8444e028ed"} Mar 14 09:29:49 crc kubenswrapper[4843]: I0314 09:29:49.195334 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" event={"ID":"eaec015a-59c1-492c-9487-003e222f49fc","Type":"ContainerStarted","Data":"85444df1f66d21898b633e04ec8c4f9ba58a50e18d75ff9c52da4a6982da2deb"} Mar 14 09:29:49 crc kubenswrapper[4843]: I0314 09:29:49.196812 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:29:49 crc kubenswrapper[4843]: I0314 09:29:49.198706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/rabbitmq-server-0" event={"ID":"ced48f61-fae8-48aa-819f-c26f7296dac5","Type":"ContainerStarted","Data":"b93013a98945b3343833c32d4c5293cc05c30d32b64018da8e4c410db866547b"} Mar 14 09:29:49 crc kubenswrapper[4843]: I0314 09:29:49.199445 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:29:49 crc kubenswrapper[4843]: I0314 09:29:49.224866 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" podStartSLOduration=38.742151232 podStartE2EDuration="57.224848576s" podCreationTimestamp="2026-03-14 09:28:52 +0000 UTC" firstStartedPulling="2026-03-14 09:28:54.595227327 +0000 UTC m=+1101.907838455" lastFinishedPulling="2026-03-14 09:29:13.077924651 +0000 UTC m=+1120.390535799" observedRunningTime="2026-03-14 09:29:49.217717888 +0000 UTC m=+1156.530329036" watchObservedRunningTime="2026-03-14 09:29:49.224848576 +0000 UTC m=+1156.537459714" Mar 14 09:29:49 crc kubenswrapper[4843]: I0314 09:29:49.247995 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/rabbitmq-server-0" podStartSLOduration=38.931531014 podStartE2EDuration="57.247975365s" podCreationTimestamp="2026-03-14 09:28:52 +0000 UTC" firstStartedPulling="2026-03-14 09:28:54.74630657 +0000 UTC m=+1102.058917698" lastFinishedPulling="2026-03-14 09:29:13.062750921 +0000 UTC m=+1120.375362049" observedRunningTime="2026-03-14 09:29:49.242958479 +0000 UTC m=+1156.555569607" watchObservedRunningTime="2026-03-14 09:29:49.247975365 +0000 UTC m=+1156.560586503" Mar 14 09:29:51 crc kubenswrapper[4843]: I0314 09:29:51.217206 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2","Type":"ContainerStarted","Data":"91aef99837eb317c628edae567c01980bc066074b43af0eb9caec32beb75da35"} Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.310086 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/root-account-create-update-mcvtm"] Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.311083 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.312999 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"openstack-mariadb-root-db-secret" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.325041 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/root-account-create-update-mcvtm"] Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.397905 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec047e38-cf88-4628-99ee-7aee2303e09f-operator-scripts\") pod \"root-account-create-update-mcvtm\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.397956 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9dlg\" (UniqueName: \"kubernetes.io/projected/ec047e38-cf88-4628-99ee-7aee2303e09f-kube-api-access-r9dlg\") pod \"root-account-create-update-mcvtm\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.499470 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec047e38-cf88-4628-99ee-7aee2303e09f-operator-scripts\") pod \"root-account-create-update-mcvtm\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.499521 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9dlg\" (UniqueName: \"kubernetes.io/projected/ec047e38-cf88-4628-99ee-7aee2303e09f-kube-api-access-r9dlg\") pod \"root-account-create-update-mcvtm\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.500439 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec047e38-cf88-4628-99ee-7aee2303e09f-operator-scripts\") pod \"root-account-create-update-mcvtm\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.519240 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9dlg\" (UniqueName: \"kubernetes.io/projected/ec047e38-cf88-4628-99ee-7aee2303e09f-kube-api-access-r9dlg\") pod \"root-account-create-update-mcvtm\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:54 crc kubenswrapper[4843]: I0314 09:29:54.630261 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.113624 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/root-account-create-update-mcvtm"] Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.251706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/root-account-create-update-mcvtm" event={"ID":"ec047e38-cf88-4628-99ee-7aee2303e09f","Type":"ContainerStarted","Data":"de899df13323c4e24c76aceb80ca264ba666a630dafd43d2356959ec0487d0ba"} Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.383947 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-db-create-bjzn8"] Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.385151 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.396461 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-create-bjzn8"] Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.493418 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc"] Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.494652 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.498834 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-db-secret" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.502730 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc"] Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.518384 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2t57\" (UniqueName: \"kubernetes.io/projected/34d732b4-52bc-42fd-9bf1-23c248b749e6-kube-api-access-g2t57\") pod \"keystone-db-create-bjzn8\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.518477 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34d732b4-52bc-42fd-9bf1-23c248b749e6-operator-scripts\") pod \"keystone-db-create-bjzn8\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.620074 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515aa0ff-df97-4090-9867-6ee2769d6750-operator-scripts\") pod \"keystone-fea7-account-create-update-ttmsc\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.620188 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34d732b4-52bc-42fd-9bf1-23c248b749e6-operator-scripts\") pod \"keystone-db-create-bjzn8\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.620265 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdwxt\" (UniqueName: \"kubernetes.io/projected/515aa0ff-df97-4090-9867-6ee2769d6750-kube-api-access-sdwxt\") pod \"keystone-fea7-account-create-update-ttmsc\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.620383 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2t57\" (UniqueName: \"kubernetes.io/projected/34d732b4-52bc-42fd-9bf1-23c248b749e6-kube-api-access-g2t57\") pod \"keystone-db-create-bjzn8\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.621149 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34d732b4-52bc-42fd-9bf1-23c248b749e6-operator-scripts\") pod \"keystone-db-create-bjzn8\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.646378 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2t57\" (UniqueName: \"kubernetes.io/projected/34d732b4-52bc-42fd-9bf1-23c248b749e6-kube-api-access-g2t57\") pod \"keystone-db-create-bjzn8\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.698474 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.722401 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdwxt\" (UniqueName: \"kubernetes.io/projected/515aa0ff-df97-4090-9867-6ee2769d6750-kube-api-access-sdwxt\") pod \"keystone-fea7-account-create-update-ttmsc\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.722536 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515aa0ff-df97-4090-9867-6ee2769d6750-operator-scripts\") pod \"keystone-fea7-account-create-update-ttmsc\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.723219 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515aa0ff-df97-4090-9867-6ee2769d6750-operator-scripts\") pod \"keystone-fea7-account-create-update-ttmsc\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.741896 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdwxt\" (UniqueName: \"kubernetes.io/projected/515aa0ff-df97-4090-9867-6ee2769d6750-kube-api-access-sdwxt\") pod \"keystone-fea7-account-create-update-ttmsc\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:55 crc kubenswrapper[4843]: I0314 09:29:55.809498 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:56 crc kubenswrapper[4843]: W0314 09:29:56.108562 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod515aa0ff_df97_4090_9867_6ee2769d6750.slice/crio-ab410be6a42ddaf6472432b25c757eab7d4b6cbe09ca4c701ed68f002e01979b WatchSource:0}: Error finding container ab410be6a42ddaf6472432b25c757eab7d4b6cbe09ca4c701ed68f002e01979b: Status 404 returned error can't find the container with id ab410be6a42ddaf6472432b25c757eab7d4b6cbe09ca4c701ed68f002e01979b Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.112123 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc"] Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.155529 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-create-bjzn8"] Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.266418 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-create-bjzn8" event={"ID":"34d732b4-52bc-42fd-9bf1-23c248b749e6","Type":"ContainerStarted","Data":"8b2bbfc15b558997b81c92b8b249c457aaf549802fb61f2d7f659c8e85dbf101"} Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.269344 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" event={"ID":"515aa0ff-df97-4090-9867-6ee2769d6750","Type":"ContainerStarted","Data":"ab410be6a42ddaf6472432b25c757eab7d4b6cbe09ca4c701ed68f002e01979b"} Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.274735 4843 generic.go:334] "Generic (PLEG): container finished" podID="ec047e38-cf88-4628-99ee-7aee2303e09f" containerID="81131ff0a422c538240e57b47ebc08dc11b257ccbecb50723ef8483f8fa8e10b" exitCode=0 Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.274799 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/root-account-create-update-mcvtm" event={"ID":"ec047e38-cf88-4628-99ee-7aee2303e09f","Type":"ContainerDied","Data":"81131ff0a422c538240e57b47ebc08dc11b257ccbecb50723ef8483f8fa8e10b"} Mar 14 09:29:56 crc kubenswrapper[4843]: I0314 09:29:56.291696 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" podStartSLOduration=1.2916782310000001 podStartE2EDuration="1.291678231s" podCreationTimestamp="2026-03-14 09:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:29:56.285561617 +0000 UTC m=+1163.598172755" watchObservedRunningTime="2026-03-14 09:29:56.291678231 +0000 UTC m=+1163.604289359" Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.294963 4843 generic.go:334] "Generic (PLEG): container finished" podID="34d732b4-52bc-42fd-9bf1-23c248b749e6" containerID="c5fe0aad2cf927dc59be42fbab76aafa7361cdbaaf9b3954d808ac182459491e" exitCode=0 Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.295022 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-create-bjzn8" event={"ID":"34d732b4-52bc-42fd-9bf1-23c248b749e6","Type":"ContainerDied","Data":"c5fe0aad2cf927dc59be42fbab76aafa7361cdbaaf9b3954d808ac182459491e"} Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.298187 4843 generic.go:334] "Generic (PLEG): container finished" podID="515aa0ff-df97-4090-9867-6ee2769d6750" containerID="c740779b5a9cfbed97f8ced9f50be3ca0d50483e67cad9c605deff98d43e1691" exitCode=0 Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.298453 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" event={"ID":"515aa0ff-df97-4090-9867-6ee2769d6750","Type":"ContainerDied","Data":"c740779b5a9cfbed97f8ced9f50be3ca0d50483e67cad9c605deff98d43e1691"} Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.695829 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.754065 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec047e38-cf88-4628-99ee-7aee2303e09f-operator-scripts\") pod \"ec047e38-cf88-4628-99ee-7aee2303e09f\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.754196 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9dlg\" (UniqueName: \"kubernetes.io/projected/ec047e38-cf88-4628-99ee-7aee2303e09f-kube-api-access-r9dlg\") pod \"ec047e38-cf88-4628-99ee-7aee2303e09f\" (UID: \"ec047e38-cf88-4628-99ee-7aee2303e09f\") " Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.755966 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec047e38-cf88-4628-99ee-7aee2303e09f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec047e38-cf88-4628-99ee-7aee2303e09f" (UID: "ec047e38-cf88-4628-99ee-7aee2303e09f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.824738 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec047e38-cf88-4628-99ee-7aee2303e09f-kube-api-access-r9dlg" (OuterVolumeSpecName: "kube-api-access-r9dlg") pod "ec047e38-cf88-4628-99ee-7aee2303e09f" (UID: "ec047e38-cf88-4628-99ee-7aee2303e09f"). InnerVolumeSpecName "kube-api-access-r9dlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.856171 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec047e38-cf88-4628-99ee-7aee2303e09f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:57 crc kubenswrapper[4843]: I0314 09:29:57.856227 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9dlg\" (UniqueName: \"kubernetes.io/projected/ec047e38-cf88-4628-99ee-7aee2303e09f-kube-api-access-r9dlg\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.307161 4843 generic.go:334] "Generic (PLEG): container finished" podID="b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2" containerID="91aef99837eb317c628edae567c01980bc066074b43af0eb9caec32beb75da35" exitCode=0 Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.307230 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2","Type":"ContainerDied","Data":"91aef99837eb317c628edae567c01980bc066074b43af0eb9caec32beb75da35"} Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.308923 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/root-account-create-update-mcvtm" event={"ID":"ec047e38-cf88-4628-99ee-7aee2303e09f","Type":"ContainerDied","Data":"de899df13323c4e24c76aceb80ca264ba666a630dafd43d2356959ec0487d0ba"} Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.308970 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de899df13323c4e24c76aceb80ca264ba666a630dafd43d2356959ec0487d0ba" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.309029 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/root-account-create-update-mcvtm" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.622138 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.652040 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.771789 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515aa0ff-df97-4090-9867-6ee2769d6750-operator-scripts\") pod \"515aa0ff-df97-4090-9867-6ee2769d6750\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.771859 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2t57\" (UniqueName: \"kubernetes.io/projected/34d732b4-52bc-42fd-9bf1-23c248b749e6-kube-api-access-g2t57\") pod \"34d732b4-52bc-42fd-9bf1-23c248b749e6\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.771953 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34d732b4-52bc-42fd-9bf1-23c248b749e6-operator-scripts\") pod \"34d732b4-52bc-42fd-9bf1-23c248b749e6\" (UID: \"34d732b4-52bc-42fd-9bf1-23c248b749e6\") " Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.771987 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdwxt\" (UniqueName: \"kubernetes.io/projected/515aa0ff-df97-4090-9867-6ee2769d6750-kube-api-access-sdwxt\") pod \"515aa0ff-df97-4090-9867-6ee2769d6750\" (UID: \"515aa0ff-df97-4090-9867-6ee2769d6750\") " Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.772640 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34d732b4-52bc-42fd-9bf1-23c248b749e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34d732b4-52bc-42fd-9bf1-23c248b749e6" (UID: "34d732b4-52bc-42fd-9bf1-23c248b749e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.772652 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/515aa0ff-df97-4090-9867-6ee2769d6750-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "515aa0ff-df97-4090-9867-6ee2769d6750" (UID: "515aa0ff-df97-4090-9867-6ee2769d6750"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.776858 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/515aa0ff-df97-4090-9867-6ee2769d6750-kube-api-access-sdwxt" (OuterVolumeSpecName: "kube-api-access-sdwxt") pod "515aa0ff-df97-4090-9867-6ee2769d6750" (UID: "515aa0ff-df97-4090-9867-6ee2769d6750"). InnerVolumeSpecName "kube-api-access-sdwxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.776959 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34d732b4-52bc-42fd-9bf1-23c248b749e6-kube-api-access-g2t57" (OuterVolumeSpecName: "kube-api-access-g2t57") pod "34d732b4-52bc-42fd-9bf1-23c248b749e6" (UID: "34d732b4-52bc-42fd-9bf1-23c248b749e6"). InnerVolumeSpecName "kube-api-access-g2t57". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.874101 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34d732b4-52bc-42fd-9bf1-23c248b749e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.874137 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdwxt\" (UniqueName: \"kubernetes.io/projected/515aa0ff-df97-4090-9867-6ee2769d6750-kube-api-access-sdwxt\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.874150 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515aa0ff-df97-4090-9867-6ee2769d6750-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:58 crc kubenswrapper[4843]: I0314 09:29:58.874161 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2t57\" (UniqueName: \"kubernetes.io/projected/34d732b4-52bc-42fd-9bf1-23c248b749e6-kube-api-access-g2t57\") on node \"crc\" DevicePath \"\"" Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.318770 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2","Type":"ContainerStarted","Data":"733fcc8d985354c1b7c9c2f396b33ad037da14fe6b6f34cabf935837ffb3ea60"} Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.320195 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-create-bjzn8" event={"ID":"34d732b4-52bc-42fd-9bf1-23c248b749e6","Type":"ContainerDied","Data":"8b2bbfc15b558997b81c92b8b249c457aaf549802fb61f2d7f659c8e85dbf101"} Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.320231 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b2bbfc15b558997b81c92b8b249c457aaf549802fb61f2d7f659c8e85dbf101" Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.320204 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-create-bjzn8" Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.321742 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" event={"ID":"515aa0ff-df97-4090-9867-6ee2769d6750","Type":"ContainerDied","Data":"ab410be6a42ddaf6472432b25c757eab7d4b6cbe09ca4c701ed68f002e01979b"} Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.321787 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab410be6a42ddaf6472432b25c757eab7d4b6cbe09ca4c701ed68f002e01979b" Mar 14 09:29:59 crc kubenswrapper[4843]: I0314 09:29:59.321765 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.136285 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558010-822zk"] Mar 14 09:30:00 crc kubenswrapper[4843]: E0314 09:30:00.136925 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="515aa0ff-df97-4090-9867-6ee2769d6750" containerName="mariadb-account-create-update" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.136937 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="515aa0ff-df97-4090-9867-6ee2769d6750" containerName="mariadb-account-create-update" Mar 14 09:30:00 crc kubenswrapper[4843]: E0314 09:30:00.136957 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec047e38-cf88-4628-99ee-7aee2303e09f" containerName="mariadb-account-create-update" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.136964 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec047e38-cf88-4628-99ee-7aee2303e09f" containerName="mariadb-account-create-update" Mar 14 09:30:00 crc kubenswrapper[4843]: E0314 09:30:00.136999 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34d732b4-52bc-42fd-9bf1-23c248b749e6" containerName="mariadb-database-create" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.137007 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="34d732b4-52bc-42fd-9bf1-23c248b749e6" containerName="mariadb-database-create" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.137157 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="515aa0ff-df97-4090-9867-6ee2769d6750" containerName="mariadb-account-create-update" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.137187 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec047e38-cf88-4628-99ee-7aee2303e09f" containerName="mariadb-account-create-update" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.137197 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="34d732b4-52bc-42fd-9bf1-23c248b749e6" containerName="mariadb-database-create" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.137813 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.140812 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.140911 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.143048 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.144321 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht"] Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.145479 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.146792 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.147074 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.157864 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558010-822zk"] Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.165181 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht"] Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.190962 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h29xd\" (UniqueName: \"kubernetes.io/projected/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b-kube-api-access-h29xd\") pod \"auto-csr-approver-29558010-822zk\" (UID: \"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b\") " pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.293081 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0fce781-5334-45d5-83be-69a184093680-secret-volume\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.293138 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msfgc\" (UniqueName: \"kubernetes.io/projected/e0fce781-5334-45d5-83be-69a184093680-kube-api-access-msfgc\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.293163 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0fce781-5334-45d5-83be-69a184093680-config-volume\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.293217 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h29xd\" (UniqueName: \"kubernetes.io/projected/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b-kube-api-access-h29xd\") pod \"auto-csr-approver-29558010-822zk\" (UID: \"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b\") " pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.310222 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h29xd\" (UniqueName: \"kubernetes.io/projected/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b-kube-api-access-h29xd\") pod \"auto-csr-approver-29558010-822zk\" (UID: \"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b\") " pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.394863 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msfgc\" (UniqueName: \"kubernetes.io/projected/e0fce781-5334-45d5-83be-69a184093680-kube-api-access-msfgc\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.394954 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0fce781-5334-45d5-83be-69a184093680-config-volume\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.395163 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0fce781-5334-45d5-83be-69a184093680-secret-volume\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.395754 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0fce781-5334-45d5-83be-69a184093680-config-volume\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.399026 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0fce781-5334-45d5-83be-69a184093680-secret-volume\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.416031 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msfgc\" (UniqueName: \"kubernetes.io/projected/e0fce781-5334-45d5-83be-69a184093680-kube-api-access-msfgc\") pod \"collect-profiles-29558010-l6rht\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.466924 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:00 crc kubenswrapper[4843]: I0314 09:30:00.480845 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.136533 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558010-822zk"] Mar 14 09:30:01 crc kubenswrapper[4843]: W0314 09:30:01.142371 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7cc93ba_b787_4d2e_9bc1_6f9d003ccd9b.slice/crio-ce45a7ab9859bd08141c0849c49cdfa23f2810d0b9c211f5db779868a98fd2c3 WatchSource:0}: Error finding container ce45a7ab9859bd08141c0849c49cdfa23f2810d0b9c211f5db779868a98fd2c3: Status 404 returned error can't find the container with id ce45a7ab9859bd08141c0849c49cdfa23f2810d0b9c211f5db779868a98fd2c3 Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.185656 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht"] Mar 14 09:30:01 crc kubenswrapper[4843]: W0314 09:30:01.187714 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0fce781_5334_45d5_83be_69a184093680.slice/crio-1761dd218535e85a9d5710c207a66ee93cfa26fb1f8b2b432611f6159676b7ff WatchSource:0}: Error finding container 1761dd218535e85a9d5710c207a66ee93cfa26fb1f8b2b432611f6159676b7ff: Status 404 returned error can't find the container with id 1761dd218535e85a9d5710c207a66ee93cfa26fb1f8b2b432611f6159676b7ff Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.351881 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558010-822zk" event={"ID":"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b","Type":"ContainerStarted","Data":"ce45a7ab9859bd08141c0849c49cdfa23f2810d0b9c211f5db779868a98fd2c3"} Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.351927 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" event={"ID":"e0fce781-5334-45d5-83be-69a184093680","Type":"ContainerStarted","Data":"d57179474278ff7904601e0c85bb6a4c87abacb33d232dcc166b8c010fff45dc"} Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.351948 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" event={"ID":"e0fce781-5334-45d5-83be-69a184093680","Type":"ContainerStarted","Data":"1761dd218535e85a9d5710c207a66ee93cfa26fb1f8b2b432611f6159676b7ff"} Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.351957 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2","Type":"ContainerStarted","Data":"81e50fbd8228d1c9e44607a928cd54a434f690bf2975c519f2e72e14bb922550"} Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.351968 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/prometheus-metric-storage-0" event={"ID":"b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2","Type":"ContainerStarted","Data":"4abe99d1de98270cd1cc260533a674f5099f1631b96b403e5ddbbeec252ecdd4"} Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.359621 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" podStartSLOduration=1.359603224 podStartE2EDuration="1.359603224s" podCreationTimestamp="2026-03-14 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:30:01.358111906 +0000 UTC m=+1168.670723044" watchObservedRunningTime="2026-03-14 09:30:01.359603224 +0000 UTC m=+1168.672214352" Mar 14 09:30:01 crc kubenswrapper[4843]: I0314 09:30:01.383738 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/prometheus-metric-storage-0" podStartSLOduration=14.383717397 podStartE2EDuration="14.383717397s" podCreationTimestamp="2026-03-14 09:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:30:01.381363558 +0000 UTC m=+1168.693974686" watchObservedRunningTime="2026-03-14 09:30:01.383717397 +0000 UTC m=+1168.696328525" Mar 14 09:30:02 crc kubenswrapper[4843]: I0314 09:30:02.367507 4843 generic.go:334] "Generic (PLEG): container finished" podID="e0fce781-5334-45d5-83be-69a184093680" containerID="d57179474278ff7904601e0c85bb6a4c87abacb33d232dcc166b8c010fff45dc" exitCode=0 Mar 14 09:30:02 crc kubenswrapper[4843]: I0314 09:30:02.368497 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" event={"ID":"e0fce781-5334-45d5-83be-69a184093680","Type":"ContainerDied","Data":"d57179474278ff7904601e0c85bb6a4c87abacb33d232dcc166b8c010fff45dc"} Mar 14 09:30:02 crc kubenswrapper[4843]: I0314 09:30:02.697245 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:30:02 crc kubenswrapper[4843]: I0314 09:30:02.697320 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:30:02 crc kubenswrapper[4843]: I0314 09:30:02.703682 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.379036 4843 generic.go:334] "Generic (PLEG): container finished" podID="c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b" containerID="73576ee29aa9d69d81deb018b6ceaab25d7626e7f004309e239637860159d359" exitCode=0 Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.379123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558010-822zk" event={"ID":"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b","Type":"ContainerDied","Data":"73576ee29aa9d69d81deb018b6ceaab25d7626e7f004309e239637860159d359"} Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.386441 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/prometheus-metric-storage-0" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.709100 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.745044 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msfgc\" (UniqueName: \"kubernetes.io/projected/e0fce781-5334-45d5-83be-69a184093680-kube-api-access-msfgc\") pod \"e0fce781-5334-45d5-83be-69a184093680\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.745118 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0fce781-5334-45d5-83be-69a184093680-config-volume\") pod \"e0fce781-5334-45d5-83be-69a184093680\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.745143 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0fce781-5334-45d5-83be-69a184093680-secret-volume\") pod \"e0fce781-5334-45d5-83be-69a184093680\" (UID: \"e0fce781-5334-45d5-83be-69a184093680\") " Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.746597 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0fce781-5334-45d5-83be-69a184093680-config-volume" (OuterVolumeSpecName: "config-volume") pod "e0fce781-5334-45d5-83be-69a184093680" (UID: "e0fce781-5334-45d5-83be-69a184093680"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.758261 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0fce781-5334-45d5-83be-69a184093680-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e0fce781-5334-45d5-83be-69a184093680" (UID: "e0fce781-5334-45d5-83be-69a184093680"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.758544 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0fce781-5334-45d5-83be-69a184093680-kube-api-access-msfgc" (OuterVolumeSpecName: "kube-api-access-msfgc") pod "e0fce781-5334-45d5-83be-69a184093680" (UID: "e0fce781-5334-45d5-83be-69a184093680"). InnerVolumeSpecName "kube-api-access-msfgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.846678 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msfgc\" (UniqueName: \"kubernetes.io/projected/e0fce781-5334-45d5-83be-69a184093680-kube-api-access-msfgc\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.846718 4843 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0fce781-5334-45d5-83be-69a184093680-config-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:03 crc kubenswrapper[4843]: I0314 09:30:03.846729 4843 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0fce781-5334-45d5-83be-69a184093680-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:04 crc kubenswrapper[4843]: I0314 09:30:04.022482 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/rabbitmq-notifications-server-0" Mar 14 09:30:04 crc kubenswrapper[4843]: I0314 09:30:04.303455 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/rabbitmq-server-0" Mar 14 09:30:04 crc kubenswrapper[4843]: I0314 09:30:04.387848 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" event={"ID":"e0fce781-5334-45d5-83be-69a184093680","Type":"ContainerDied","Data":"1761dd218535e85a9d5710c207a66ee93cfa26fb1f8b2b432611f6159676b7ff"} Mar 14 09:30:04 crc kubenswrapper[4843]: I0314 09:30:04.387949 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1761dd218535e85a9d5710c207a66ee93cfa26fb1f8b2b432611f6159676b7ff" Mar 14 09:30:04 crc kubenswrapper[4843]: I0314 09:30:04.388012 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558010-l6rht" Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:04.698822 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:04.762147 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h29xd\" (UniqueName: \"kubernetes.io/projected/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b-kube-api-access-h29xd\") pod \"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b\" (UID: \"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b\") " Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:04.767310 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b-kube-api-access-h29xd" (OuterVolumeSpecName: "kube-api-access-h29xd") pod "c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b" (UID: "c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b"). InnerVolumeSpecName "kube-api-access-h29xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:04.863988 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h29xd\" (UniqueName: \"kubernetes.io/projected/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b-kube-api-access-h29xd\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:05.400510 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558010-822zk" event={"ID":"c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b","Type":"ContainerDied","Data":"ce45a7ab9859bd08141c0849c49cdfa23f2810d0b9c211f5db779868a98fd2c3"} Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:05.400873 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce45a7ab9859bd08141c0849c49cdfa23f2810d0b9c211f5db779868a98fd2c3" Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:05.400976 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558010-822zk" Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:05.791828 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558004-jlkwf"] Mar 14 09:30:05 crc kubenswrapper[4843]: I0314 09:30:05.800529 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558004-jlkwf"] Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.094834 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kb65g"] Mar 14 09:30:06 crc kubenswrapper[4843]: E0314 09:30:06.095168 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b" containerName="oc" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.095187 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b" containerName="oc" Mar 14 09:30:06 crc kubenswrapper[4843]: E0314 09:30:06.095206 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fce781-5334-45d5-83be-69a184093680" containerName="collect-profiles" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.095215 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fce781-5334-45d5-83be-69a184093680" containerName="collect-profiles" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.095407 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0fce781-5334-45d5-83be-69a184093680" containerName="collect-profiles" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.095428 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b" containerName="oc" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.095938 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.097496 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.098859 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-tw4tg" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.101164 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.101474 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.105483 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kb65g"] Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.282764 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-config-data\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.282847 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-combined-ca-bundle\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.282878 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzzks\" (UniqueName: \"kubernetes.io/projected/8d99019e-ec8b-4310-92c5-a764541ac1ce-kube-api-access-zzzks\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.384566 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-combined-ca-bundle\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.384642 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzzks\" (UniqueName: \"kubernetes.io/projected/8d99019e-ec8b-4310-92c5-a764541ac1ce-kube-api-access-zzzks\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.384768 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-config-data\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.389996 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-combined-ca-bundle\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.402211 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-config-data\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.422716 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzzks\" (UniqueName: \"kubernetes.io/projected/8d99019e-ec8b-4310-92c5-a764541ac1ce-kube-api-access-zzzks\") pod \"keystone-db-sync-kb65g\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.709967 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:06 crc kubenswrapper[4843]: I0314 09:30:06.966921 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kb65g"] Mar 14 09:30:06 crc kubenswrapper[4843]: W0314 09:30:06.970266 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d99019e_ec8b_4310_92c5_a764541ac1ce.slice/crio-20e541da64e93e4246538e94d4edb065754fdba524dc1d859eeb447611bb9b4d WatchSource:0}: Error finding container 20e541da64e93e4246538e94d4edb065754fdba524dc1d859eeb447611bb9b4d: Status 404 returned error can't find the container with id 20e541da64e93e4246538e94d4edb065754fdba524dc1d859eeb447611bb9b4d Mar 14 09:30:07 crc kubenswrapper[4843]: I0314 09:30:07.349527 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ff20372-3c55-4465-8ddf-8467007709a3" path="/var/lib/kubelet/pods/8ff20372-3c55-4465-8ddf-8467007709a3/volumes" Mar 14 09:30:07 crc kubenswrapper[4843]: I0314 09:30:07.417454 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kb65g" event={"ID":"8d99019e-ec8b-4310-92c5-a764541ac1ce","Type":"ContainerStarted","Data":"20e541da64e93e4246538e94d4edb065754fdba524dc1d859eeb447611bb9b4d"} Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.387776 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.388345 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.388396 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.389095 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c0b6bcfd5822f50fc94ed067fd20658758c736e480b5fd0ec55da8213db2e74"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.389159 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://1c0b6bcfd5822f50fc94ed067fd20658758c736e480b5fd0ec55da8213db2e74" gracePeriod=600 Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.484582 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kb65g" event={"ID":"8d99019e-ec8b-4310-92c5-a764541ac1ce","Type":"ContainerStarted","Data":"5b598fdfa71178c7ea0d960f8aaf659cf2e45cd1bac4489cdb34f277a86591fb"} Mar 14 09:30:15 crc kubenswrapper[4843]: I0314 09:30:15.531341 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-db-sync-kb65g" podStartSLOduration=2.156407315 podStartE2EDuration="9.531324864s" podCreationTimestamp="2026-03-14 09:30:06 +0000 UTC" firstStartedPulling="2026-03-14 09:30:06.972051345 +0000 UTC m=+1174.284662473" lastFinishedPulling="2026-03-14 09:30:14.346968884 +0000 UTC m=+1181.659580022" observedRunningTime="2026-03-14 09:30:15.530427761 +0000 UTC m=+1182.843038889" watchObservedRunningTime="2026-03-14 09:30:15.531324864 +0000 UTC m=+1182.843935992" Mar 14 09:30:16 crc kubenswrapper[4843]: I0314 09:30:16.498118 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="1c0b6bcfd5822f50fc94ed067fd20658758c736e480b5fd0ec55da8213db2e74" exitCode=0 Mar 14 09:30:16 crc kubenswrapper[4843]: I0314 09:30:16.498232 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"1c0b6bcfd5822f50fc94ed067fd20658758c736e480b5fd0ec55da8213db2e74"} Mar 14 09:30:16 crc kubenswrapper[4843]: I0314 09:30:16.498495 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"ea69e0d05b7309b15d93d55c71a74bebdf19a7967db0d10155ee8a1a048179ea"} Mar 14 09:30:16 crc kubenswrapper[4843]: I0314 09:30:16.498520 4843 scope.go:117] "RemoveContainer" containerID="144fcfbfce69a65758af21b0b25db69c4308200478b4f5a2c753f3696a587b2b" Mar 14 09:30:17 crc kubenswrapper[4843]: I0314 09:30:17.512734 4843 generic.go:334] "Generic (PLEG): container finished" podID="8d99019e-ec8b-4310-92c5-a764541ac1ce" containerID="5b598fdfa71178c7ea0d960f8aaf659cf2e45cd1bac4489cdb34f277a86591fb" exitCode=0 Mar 14 09:30:17 crc kubenswrapper[4843]: I0314 09:30:17.512861 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kb65g" event={"ID":"8d99019e-ec8b-4310-92c5-a764541ac1ce","Type":"ContainerDied","Data":"5b598fdfa71178c7ea0d960f8aaf659cf2e45cd1bac4489cdb34f277a86591fb"} Mar 14 09:30:18 crc kubenswrapper[4843]: I0314 09:30:18.873550 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:18 crc kubenswrapper[4843]: I0314 09:30:18.989950 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-config-data\") pod \"8d99019e-ec8b-4310-92c5-a764541ac1ce\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " Mar 14 09:30:18 crc kubenswrapper[4843]: I0314 09:30:18.990152 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzzks\" (UniqueName: \"kubernetes.io/projected/8d99019e-ec8b-4310-92c5-a764541ac1ce-kube-api-access-zzzks\") pod \"8d99019e-ec8b-4310-92c5-a764541ac1ce\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " Mar 14 09:30:18 crc kubenswrapper[4843]: I0314 09:30:18.990189 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-combined-ca-bundle\") pod \"8d99019e-ec8b-4310-92c5-a764541ac1ce\" (UID: \"8d99019e-ec8b-4310-92c5-a764541ac1ce\") " Mar 14 09:30:18 crc kubenswrapper[4843]: I0314 09:30:18.995401 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d99019e-ec8b-4310-92c5-a764541ac1ce-kube-api-access-zzzks" (OuterVolumeSpecName: "kube-api-access-zzzks") pod "8d99019e-ec8b-4310-92c5-a764541ac1ce" (UID: "8d99019e-ec8b-4310-92c5-a764541ac1ce"). InnerVolumeSpecName "kube-api-access-zzzks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.012129 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d99019e-ec8b-4310-92c5-a764541ac1ce" (UID: "8d99019e-ec8b-4310-92c5-a764541ac1ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.034579 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-config-data" (OuterVolumeSpecName: "config-data") pod "8d99019e-ec8b-4310-92c5-a764541ac1ce" (UID: "8d99019e-ec8b-4310-92c5-a764541ac1ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.092023 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.092069 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzzks\" (UniqueName: \"kubernetes.io/projected/8d99019e-ec8b-4310-92c5-a764541ac1ce-kube-api-access-zzzks\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.092085 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d99019e-ec8b-4310-92c5-a764541ac1ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.529202 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-db-sync-kb65g" event={"ID":"8d99019e-ec8b-4310-92c5-a764541ac1ce","Type":"ContainerDied","Data":"20e541da64e93e4246538e94d4edb065754fdba524dc1d859eeb447611bb9b4d"} Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.529236 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-db-sync-kb65g" Mar 14 09:30:19 crc kubenswrapper[4843]: I0314 09:30:19.529244 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20e541da64e93e4246538e94d4edb065754fdba524dc1d859eeb447611bb9b4d" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.089079 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-r2kxf"] Mar 14 09:30:20 crc kubenswrapper[4843]: E0314 09:30:20.089770 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d99019e-ec8b-4310-92c5-a764541ac1ce" containerName="keystone-db-sync" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.089785 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d99019e-ec8b-4310-92c5-a764541ac1ce" containerName="keystone-db-sync" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.089978 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d99019e-ec8b-4310-92c5-a764541ac1ce" containerName="keystone-db-sync" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.091496 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.094593 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"osp-secret" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.094700 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.094763 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.094834 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-tw4tg" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.094989 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.102669 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-r2kxf"] Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.209157 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-config-data\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.209197 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-fernet-keys\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.209226 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-credential-keys\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.209265 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8vm5\" (UniqueName: \"kubernetes.io/projected/94608b81-49f2-43a4-9af6-3069fb9f71ca-kube-api-access-z8vm5\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.209343 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-scripts\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.209471 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-combined-ca-bundle\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.231953 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.234413 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.237933 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.242860 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.244762 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.311051 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-config-data\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.311116 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-fernet-keys\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.311151 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-credential-keys\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.311203 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8vm5\" (UniqueName: \"kubernetes.io/projected/94608b81-49f2-43a4-9af6-3069fb9f71ca-kube-api-access-z8vm5\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.311244 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-scripts\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.311314 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-combined-ca-bundle\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.317171 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-config-data\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.319025 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-scripts\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.319828 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-fernet-keys\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.321011 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-combined-ca-bundle\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.327990 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-credential-keys\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.340565 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8vm5\" (UniqueName: \"kubernetes.io/projected/94608b81-49f2-43a4-9af6-3069fb9f71ca-kube-api-access-z8vm5\") pod \"keystone-bootstrap-r2kxf\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.407164 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.412893 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-scripts\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.412946 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-log-httpd\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.413000 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-config-data\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.413058 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.413114 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.413142 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-run-httpd\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.413165 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm5xb\" (UniqueName: \"kubernetes.io/projected/0c4c6c00-6874-4581-8123-8f0e506626a5-kube-api-access-cm5xb\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517310 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-log-httpd\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517495 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-config-data\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517574 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517676 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517701 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-run-httpd\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517727 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm5xb\" (UniqueName: \"kubernetes.io/projected/0c4c6c00-6874-4581-8123-8f0e506626a5-kube-api-access-cm5xb\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.517806 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-scripts\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.519986 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-log-httpd\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.522152 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-run-httpd\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.541126 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.541309 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.548577 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-scripts\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.551323 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm5xb\" (UniqueName: \"kubernetes.io/projected/0c4c6c00-6874-4581-8123-8f0e506626a5-kube-api-access-cm5xb\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.554489 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-config-data\") pod \"ceilometer-0\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.851628 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:20 crc kubenswrapper[4843]: I0314 09:30:20.865031 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-r2kxf"] Mar 14 09:30:20 crc kubenswrapper[4843]: W0314 09:30:20.876067 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94608b81_49f2_43a4_9af6_3069fb9f71ca.slice/crio-a6233efb7227ac949e05eaa374049fee20562d6302834aff154c03dffb47b823 WatchSource:0}: Error finding container a6233efb7227ac949e05eaa374049fee20562d6302834aff154c03dffb47b823: Status 404 returned error can't find the container with id a6233efb7227ac949e05eaa374049fee20562d6302834aff154c03dffb47b823 Mar 14 09:30:21 crc kubenswrapper[4843]: I0314 09:30:21.376432 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:21 crc kubenswrapper[4843]: I0314 09:30:21.559262 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" event={"ID":"94608b81-49f2-43a4-9af6-3069fb9f71ca","Type":"ContainerStarted","Data":"60483461e6971b935c938888e01f74fbff2ad70a71ad49257931eae9c5033cc7"} Mar 14 09:30:21 crc kubenswrapper[4843]: I0314 09:30:21.559573 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" event={"ID":"94608b81-49f2-43a4-9af6-3069fb9f71ca","Type":"ContainerStarted","Data":"a6233efb7227ac949e05eaa374049fee20562d6302834aff154c03dffb47b823"} Mar 14 09:30:21 crc kubenswrapper[4843]: I0314 09:30:21.561582 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerStarted","Data":"a273cca5e9f4636356e8dc83250cd8e4bd9e4ed3345a0551a11861225912b787"} Mar 14 09:30:21 crc kubenswrapper[4843]: I0314 09:30:21.577771 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" podStartSLOduration=1.577752178 podStartE2EDuration="1.577752178s" podCreationTimestamp="2026-03-14 09:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:30:21.575221475 +0000 UTC m=+1188.887832623" watchObservedRunningTime="2026-03-14 09:30:21.577752178 +0000 UTC m=+1188.890363296" Mar 14 09:30:22 crc kubenswrapper[4843]: I0314 09:30:22.110165 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:24 crc kubenswrapper[4843]: I0314 09:30:24.605434 4843 generic.go:334] "Generic (PLEG): container finished" podID="94608b81-49f2-43a4-9af6-3069fb9f71ca" containerID="60483461e6971b935c938888e01f74fbff2ad70a71ad49257931eae9c5033cc7" exitCode=0 Mar 14 09:30:24 crc kubenswrapper[4843]: I0314 09:30:24.605493 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" event={"ID":"94608b81-49f2-43a4-9af6-3069fb9f71ca","Type":"ContainerDied","Data":"60483461e6971b935c938888e01f74fbff2ad70a71ad49257931eae9c5033cc7"} Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.078856 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.221104 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-credential-keys\") pod \"94608b81-49f2-43a4-9af6-3069fb9f71ca\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.222145 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8vm5\" (UniqueName: \"kubernetes.io/projected/94608b81-49f2-43a4-9af6-3069fb9f71ca-kube-api-access-z8vm5\") pod \"94608b81-49f2-43a4-9af6-3069fb9f71ca\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.222247 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-config-data\") pod \"94608b81-49f2-43a4-9af6-3069fb9f71ca\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.222386 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-scripts\") pod \"94608b81-49f2-43a4-9af6-3069fb9f71ca\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.222444 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-fernet-keys\") pod \"94608b81-49f2-43a4-9af6-3069fb9f71ca\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.222478 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-combined-ca-bundle\") pod \"94608b81-49f2-43a4-9af6-3069fb9f71ca\" (UID: \"94608b81-49f2-43a4-9af6-3069fb9f71ca\") " Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.225734 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "94608b81-49f2-43a4-9af6-3069fb9f71ca" (UID: "94608b81-49f2-43a4-9af6-3069fb9f71ca"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.226087 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "94608b81-49f2-43a4-9af6-3069fb9f71ca" (UID: "94608b81-49f2-43a4-9af6-3069fb9f71ca"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.226120 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-scripts" (OuterVolumeSpecName: "scripts") pod "94608b81-49f2-43a4-9af6-3069fb9f71ca" (UID: "94608b81-49f2-43a4-9af6-3069fb9f71ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.226235 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94608b81-49f2-43a4-9af6-3069fb9f71ca-kube-api-access-z8vm5" (OuterVolumeSpecName: "kube-api-access-z8vm5") pod "94608b81-49f2-43a4-9af6-3069fb9f71ca" (UID: "94608b81-49f2-43a4-9af6-3069fb9f71ca"). InnerVolumeSpecName "kube-api-access-z8vm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.247698 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-config-data" (OuterVolumeSpecName: "config-data") pod "94608b81-49f2-43a4-9af6-3069fb9f71ca" (UID: "94608b81-49f2-43a4-9af6-3069fb9f71ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.249833 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94608b81-49f2-43a4-9af6-3069fb9f71ca" (UID: "94608b81-49f2-43a4-9af6-3069fb9f71ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.324750 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8vm5\" (UniqueName: \"kubernetes.io/projected/94608b81-49f2-43a4-9af6-3069fb9f71ca-kube-api-access-z8vm5\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.324808 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.324830 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.324849 4843 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.324866 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.324884 4843 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94608b81-49f2-43a4-9af6-3069fb9f71ca-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.622874 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" event={"ID":"94608b81-49f2-43a4-9af6-3069fb9f71ca","Type":"ContainerDied","Data":"a6233efb7227ac949e05eaa374049fee20562d6302834aff154c03dffb47b823"} Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.622930 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6233efb7227ac949e05eaa374049fee20562d6302834aff154c03dffb47b823" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.622889 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-r2kxf" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.624706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerStarted","Data":"94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c"} Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.719567 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-r2kxf"] Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.726995 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-r2kxf"] Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.815254 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-cmr6j"] Mar 14 09:30:26 crc kubenswrapper[4843]: E0314 09:30:26.815634 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94608b81-49f2-43a4-9af6-3069fb9f71ca" containerName="keystone-bootstrap" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.815650 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="94608b81-49f2-43a4-9af6-3069fb9f71ca" containerName="keystone-bootstrap" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.815794 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="94608b81-49f2-43a4-9af6-3069fb9f71ca" containerName="keystone-bootstrap" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.816317 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.817844 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.818644 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.818913 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-tw4tg" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.819427 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"osp-secret" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.819579 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.830947 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-cmr6j"] Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.937300 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-config-data\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.937425 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-fernet-keys\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.937480 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-credential-keys\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.937582 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22k44\" (UniqueName: \"kubernetes.io/projected/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-kube-api-access-22k44\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.937645 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-combined-ca-bundle\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:26 crc kubenswrapper[4843]: I0314 09:30:26.937776 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-scripts\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.044424 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22k44\" (UniqueName: \"kubernetes.io/projected/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-kube-api-access-22k44\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.044520 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-combined-ca-bundle\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.044589 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-scripts\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.044663 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-config-data\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.044707 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-fernet-keys\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.048394 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-credential-keys\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.053357 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-credential-keys\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.057062 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-config-data\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.057371 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-scripts\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.057624 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-combined-ca-bundle\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.061163 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-fernet-keys\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.073676 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22k44\" (UniqueName: \"kubernetes.io/projected/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-kube-api-access-22k44\") pod \"keystone-bootstrap-cmr6j\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.134799 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.354539 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94608b81-49f2-43a4-9af6-3069fb9f71ca" path="/var/lib/kubelet/pods/94608b81-49f2-43a4-9af6-3069fb9f71ca/volumes" Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.598442 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-cmr6j"] Mar 14 09:30:27 crc kubenswrapper[4843]: W0314 09:30:27.605581 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ca31b1f_982b_44fe_977d_6e2c9d2d007f.slice/crio-df66c4b84b9790fc17e3206b3d43b93c75e5aefb5584d8e42610173b155a194d WatchSource:0}: Error finding container df66c4b84b9790fc17e3206b3d43b93c75e5aefb5584d8e42610173b155a194d: Status 404 returned error can't find the container with id df66c4b84b9790fc17e3206b3d43b93c75e5aefb5584d8e42610173b155a194d Mar 14 09:30:27 crc kubenswrapper[4843]: I0314 09:30:27.636746 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" event={"ID":"7ca31b1f-982b-44fe-977d-6e2c9d2d007f","Type":"ContainerStarted","Data":"df66c4b84b9790fc17e3206b3d43b93c75e5aefb5584d8e42610173b155a194d"} Mar 14 09:30:28 crc kubenswrapper[4843]: I0314 09:30:28.646433 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerStarted","Data":"fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c"} Mar 14 09:30:28 crc kubenswrapper[4843]: I0314 09:30:28.648323 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" event={"ID":"7ca31b1f-982b-44fe-977d-6e2c9d2d007f","Type":"ContainerStarted","Data":"3466ed6f81b7920e04e49be9c7db03edd9d3c4c4b7b28d209a022100a393cf79"} Mar 14 09:30:28 crc kubenswrapper[4843]: I0314 09:30:28.670027 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" podStartSLOduration=2.669811103 podStartE2EDuration="2.669811103s" podCreationTimestamp="2026-03-14 09:30:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:30:28.666385238 +0000 UTC m=+1195.978996376" watchObservedRunningTime="2026-03-14 09:30:28.669811103 +0000 UTC m=+1195.982422231" Mar 14 09:30:31 crc kubenswrapper[4843]: I0314 09:30:31.680927 4843 generic.go:334] "Generic (PLEG): container finished" podID="7ca31b1f-982b-44fe-977d-6e2c9d2d007f" containerID="3466ed6f81b7920e04e49be9c7db03edd9d3c4c4b7b28d209a022100a393cf79" exitCode=0 Mar 14 09:30:31 crc kubenswrapper[4843]: I0314 09:30:31.681008 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" event={"ID":"7ca31b1f-982b-44fe-977d-6e2c9d2d007f","Type":"ContainerDied","Data":"3466ed6f81b7920e04e49be9c7db03edd9d3c4c4b7b28d209a022100a393cf79"} Mar 14 09:30:32 crc kubenswrapper[4843]: I0314 09:30:32.690491 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerStarted","Data":"6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d"} Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.025034 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.081423 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22k44\" (UniqueName: \"kubernetes.io/projected/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-kube-api-access-22k44\") pod \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.081476 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-config-data\") pod \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.081528 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-combined-ca-bundle\") pod \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.081555 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-scripts\") pod \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.081642 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-credential-keys\") pod \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.081669 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-fernet-keys\") pod \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\" (UID: \"7ca31b1f-982b-44fe-977d-6e2c9d2d007f\") " Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.117727 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7ca31b1f-982b-44fe-977d-6e2c9d2d007f" (UID: "7ca31b1f-982b-44fe-977d-6e2c9d2d007f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.124996 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-kube-api-access-22k44" (OuterVolumeSpecName: "kube-api-access-22k44") pod "7ca31b1f-982b-44fe-977d-6e2c9d2d007f" (UID: "7ca31b1f-982b-44fe-977d-6e2c9d2d007f"). InnerVolumeSpecName "kube-api-access-22k44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.125138 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7ca31b1f-982b-44fe-977d-6e2c9d2d007f" (UID: "7ca31b1f-982b-44fe-977d-6e2c9d2d007f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.126561 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-scripts" (OuterVolumeSpecName: "scripts") pod "7ca31b1f-982b-44fe-977d-6e2c9d2d007f" (UID: "7ca31b1f-982b-44fe-977d-6e2c9d2d007f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.129823 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-config-data" (OuterVolumeSpecName: "config-data") pod "7ca31b1f-982b-44fe-977d-6e2c9d2d007f" (UID: "7ca31b1f-982b-44fe-977d-6e2c9d2d007f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.130005 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ca31b1f-982b-44fe-977d-6e2c9d2d007f" (UID: "7ca31b1f-982b-44fe-977d-6e2c9d2d007f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.189514 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.189551 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.189564 4843 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.189574 4843 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.189585 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22k44\" (UniqueName: \"kubernetes.io/projected/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-kube-api-access-22k44\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.189595 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ca31b1f-982b-44fe-977d-6e2c9d2d007f-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.727201 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" event={"ID":"7ca31b1f-982b-44fe-977d-6e2c9d2d007f","Type":"ContainerDied","Data":"df66c4b84b9790fc17e3206b3d43b93c75e5aefb5584d8e42610173b155a194d"} Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.728222 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df66c4b84b9790fc17e3206b3d43b93c75e5aefb5584d8e42610173b155a194d" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.727384 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-cmr6j" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.785011 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-577555cd46-j7w69"] Mar 14 09:30:33 crc kubenswrapper[4843]: E0314 09:30:33.785356 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca31b1f-982b-44fe-977d-6e2c9d2d007f" containerName="keystone-bootstrap" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.785373 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca31b1f-982b-44fe-977d-6e2c9d2d007f" containerName="keystone-bootstrap" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.785524 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ca31b1f-982b-44fe-977d-6e2c9d2d007f" containerName="keystone-bootstrap" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.786012 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.789587 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-keystone-internal-svc" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.789608 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-keystone-dockercfg-tw4tg" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.789722 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-keystone-public-svc" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.789866 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-config-data" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.791071 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.796361 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"keystone-scripts" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.802499 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-577555cd46-j7w69"] Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.920954 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vpcl\" (UniqueName: \"kubernetes.io/projected/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-kube-api-access-6vpcl\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.920989 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-internal-tls-certs\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.921019 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-config-data\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.921037 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-credential-keys\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.921057 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-scripts\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.921098 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-public-tls-certs\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.921126 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-fernet-keys\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:33 crc kubenswrapper[4843]: I0314 09:30:33.921145 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-combined-ca-bundle\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025227 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vpcl\" (UniqueName: \"kubernetes.io/projected/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-kube-api-access-6vpcl\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025348 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-internal-tls-certs\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025424 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-config-data\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025490 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-credential-keys\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025556 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-scripts\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025649 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-public-tls-certs\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025718 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-fernet-keys\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.025784 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-combined-ca-bundle\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.035891 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-internal-tls-certs\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.035955 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-scripts\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.036464 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-public-tls-certs\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.038092 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-config-data\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.038106 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-fernet-keys\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.038737 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-credential-keys\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.044015 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-combined-ca-bundle\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.045029 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vpcl\" (UniqueName: \"kubernetes.io/projected/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-kube-api-access-6vpcl\") pod \"keystone-577555cd46-j7w69\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.113105 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.556625 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-577555cd46-j7w69"] Mar 14 09:30:34 crc kubenswrapper[4843]: W0314 09:30:34.564802 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb9f076c_8aa4_41c0_aa84_6a1eb0239a86.slice/crio-4b8e10f76dc986698b74dcad7d5529c7a7cb5d40edcfc206354ac03cc45de422 WatchSource:0}: Error finding container 4b8e10f76dc986698b74dcad7d5529c7a7cb5d40edcfc206354ac03cc45de422: Status 404 returned error can't find the container with id 4b8e10f76dc986698b74dcad7d5529c7a7cb5d40edcfc206354ac03cc45de422 Mar 14 09:30:34 crc kubenswrapper[4843]: I0314 09:30:34.741815 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" event={"ID":"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86","Type":"ContainerStarted","Data":"4b8e10f76dc986698b74dcad7d5529c7a7cb5d40edcfc206354ac03cc45de422"} Mar 14 09:30:35 crc kubenswrapper[4843]: I0314 09:30:35.752800 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" event={"ID":"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86","Type":"ContainerStarted","Data":"80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8"} Mar 14 09:30:35 crc kubenswrapper[4843]: I0314 09:30:35.753088 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:30:35 crc kubenswrapper[4843]: I0314 09:30:35.771298 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" podStartSLOduration=2.771265102 podStartE2EDuration="2.771265102s" podCreationTimestamp="2026-03-14 09:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:30:35.769315994 +0000 UTC m=+1203.081927122" watchObservedRunningTime="2026-03-14 09:30:35.771265102 +0000 UTC m=+1203.083876230" Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.863473 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerStarted","Data":"1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7"} Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.864418 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.863722 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-notification-agent" containerID="cri-o://fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c" gracePeriod=30 Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.863657 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-central-agent" containerID="cri-o://94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c" gracePeriod=30 Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.863727 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="proxy-httpd" containerID="cri-o://1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7" gracePeriod=30 Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.863751 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="sg-core" containerID="cri-o://6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d" gracePeriod=30 Mar 14 09:30:42 crc kubenswrapper[4843]: I0314 09:30:42.894070 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.074879542 podStartE2EDuration="22.894051036s" podCreationTimestamp="2026-03-14 09:30:20 +0000 UTC" firstStartedPulling="2026-03-14 09:30:21.385309096 +0000 UTC m=+1188.697920224" lastFinishedPulling="2026-03-14 09:30:42.20448058 +0000 UTC m=+1209.517091718" observedRunningTime="2026-03-14 09:30:42.892869336 +0000 UTC m=+1210.205480504" watchObservedRunningTime="2026-03-14 09:30:42.894051036 +0000 UTC m=+1210.206662154" Mar 14 09:30:43 crc kubenswrapper[4843]: I0314 09:30:43.873978 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerID="1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7" exitCode=0 Mar 14 09:30:43 crc kubenswrapper[4843]: I0314 09:30:43.874240 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerID="6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d" exitCode=2 Mar 14 09:30:43 crc kubenswrapper[4843]: I0314 09:30:43.874251 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerID="94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c" exitCode=0 Mar 14 09:30:43 crc kubenswrapper[4843]: I0314 09:30:43.874052 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerDied","Data":"1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7"} Mar 14 09:30:43 crc kubenswrapper[4843]: I0314 09:30:43.874303 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerDied","Data":"6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d"} Mar 14 09:30:43 crc kubenswrapper[4843]: I0314 09:30:43.874324 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerDied","Data":"94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c"} Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.114538 4843 scope.go:117] "RemoveContainer" containerID="3ff8c1547f4171613d9750fb339269d2ecf0f9df9397830da34db78298934b49" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.782572 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854546 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-sg-core-conf-yaml\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854597 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-scripts\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854663 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-combined-ca-bundle\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854691 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-log-httpd\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854717 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-run-httpd\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854745 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-config-data\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.854780 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm5xb\" (UniqueName: \"kubernetes.io/projected/0c4c6c00-6874-4581-8123-8f0e506626a5-kube-api-access-cm5xb\") pod \"0c4c6c00-6874-4581-8123-8f0e506626a5\" (UID: \"0c4c6c00-6874-4581-8123-8f0e506626a5\") " Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.857630 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.858176 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.861757 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c4c6c00-6874-4581-8123-8f0e506626a5-kube-api-access-cm5xb" (OuterVolumeSpecName: "kube-api-access-cm5xb") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "kube-api-access-cm5xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.865795 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-scripts" (OuterVolumeSpecName: "scripts") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.880368 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.893787 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerID="fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c" exitCode=0 Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.893843 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.893865 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerDied","Data":"fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c"} Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.893908 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0c4c6c00-6874-4581-8123-8f0e506626a5","Type":"ContainerDied","Data":"a273cca5e9f4636356e8dc83250cd8e4bd9e4ed3345a0551a11861225912b787"} Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.893928 4843 scope.go:117] "RemoveContainer" containerID="1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.913815 4843 scope.go:117] "RemoveContainer" containerID="6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.921216 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.928331 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-config-data" (OuterVolumeSpecName: "config-data") pod "0c4c6c00-6874-4581-8123-8f0e506626a5" (UID: "0c4c6c00-6874-4581-8123-8f0e506626a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.932046 4843 scope.go:117] "RemoveContainer" containerID="fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.947983 4843 scope.go:117] "RemoveContainer" containerID="94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956445 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956468 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956477 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956488 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956497 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c4c6c00-6874-4581-8123-8f0e506626a5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956504 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c4c6c00-6874-4581-8123-8f0e506626a5-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.956512 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm5xb\" (UniqueName: \"kubernetes.io/projected/0c4c6c00-6874-4581-8123-8f0e506626a5-kube-api-access-cm5xb\") on node \"crc\" DevicePath \"\"" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.973627 4843 scope.go:117] "RemoveContainer" containerID="1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7" Mar 14 09:30:45 crc kubenswrapper[4843]: E0314 09:30:45.974074 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7\": container with ID starting with 1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7 not found: ID does not exist" containerID="1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.974105 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7"} err="failed to get container status \"1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7\": rpc error: code = NotFound desc = could not find container \"1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7\": container with ID starting with 1bd5fdc44a92e9f2d34381e35d441b6674bad50db3fb5c15829e2d227fd1dec7 not found: ID does not exist" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.974126 4843 scope.go:117] "RemoveContainer" containerID="6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d" Mar 14 09:30:45 crc kubenswrapper[4843]: E0314 09:30:45.974458 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d\": container with ID starting with 6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d not found: ID does not exist" containerID="6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.974477 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d"} err="failed to get container status \"6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d\": rpc error: code = NotFound desc = could not find container \"6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d\": container with ID starting with 6b468fe809d1fdc2df21c1686da44136e8fa43ebc0199d9bb033e34a57bf0b2d not found: ID does not exist" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.974490 4843 scope.go:117] "RemoveContainer" containerID="fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c" Mar 14 09:30:45 crc kubenswrapper[4843]: E0314 09:30:45.974758 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c\": container with ID starting with fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c not found: ID does not exist" containerID="fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.974801 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c"} err="failed to get container status \"fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c\": rpc error: code = NotFound desc = could not find container \"fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c\": container with ID starting with fae08128870be5ba91cf65d2ea1286ce6a1ffdaba8485af44ba9d7d5e3b9e69c not found: ID does not exist" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.974831 4843 scope.go:117] "RemoveContainer" containerID="94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c" Mar 14 09:30:45 crc kubenswrapper[4843]: E0314 09:30:45.975139 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c\": container with ID starting with 94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c not found: ID does not exist" containerID="94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c" Mar 14 09:30:45 crc kubenswrapper[4843]: I0314 09:30:45.975173 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c"} err="failed to get container status \"94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c\": rpc error: code = NotFound desc = could not find container \"94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c\": container with ID starting with 94c392ffd0bbab2f4c27019e116469d34feaeb53fc7e4467fd9cde9db356070c not found: ID does not exist" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.229674 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.236342 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.256837 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:46 crc kubenswrapper[4843]: E0314 09:30:46.257119 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-notification-agent" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257132 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-notification-agent" Mar 14 09:30:46 crc kubenswrapper[4843]: E0314 09:30:46.257153 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="sg-core" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257161 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="sg-core" Mar 14 09:30:46 crc kubenswrapper[4843]: E0314 09:30:46.257171 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-central-agent" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257177 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-central-agent" Mar 14 09:30:46 crc kubenswrapper[4843]: E0314 09:30:46.257187 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="proxy-httpd" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257194 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="proxy-httpd" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257374 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-notification-agent" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257392 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="ceilometer-central-agent" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257403 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="proxy-httpd" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.257412 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" containerName="sg-core" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.258797 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.261835 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.261835 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.274523 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362385 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-log-httpd\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362468 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl5wx\" (UniqueName: \"kubernetes.io/projected/008369ae-6b4f-4dbf-ab72-d279508616b7-kube-api-access-zl5wx\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362499 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-config-data\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362530 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362613 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-run-httpd\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362667 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-scripts\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.362710 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.464836 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.464932 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-run-httpd\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465013 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-scripts\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465076 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465192 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-log-httpd\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465230 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl5wx\" (UniqueName: \"kubernetes.io/projected/008369ae-6b4f-4dbf-ab72-d279508616b7-kube-api-access-zl5wx\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465315 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-config-data\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465571 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-log-httpd\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.465762 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-run-httpd\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.468968 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.468995 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-scripts\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.474188 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-config-data\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.474637 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.490050 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl5wx\" (UniqueName: \"kubernetes.io/projected/008369ae-6b4f-4dbf-ab72-d279508616b7-kube-api-access-zl5wx\") pod \"ceilometer-0\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.575176 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:46 crc kubenswrapper[4843]: I0314 09:30:46.999078 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:30:47 crc kubenswrapper[4843]: W0314 09:30:47.008102 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008369ae_6b4f_4dbf_ab72_d279508616b7.slice/crio-53c47655c854f932df1376117113e85cb1b632b0878a0516fcadade01e805b01 WatchSource:0}: Error finding container 53c47655c854f932df1376117113e85cb1b632b0878a0516fcadade01e805b01: Status 404 returned error can't find the container with id 53c47655c854f932df1376117113e85cb1b632b0878a0516fcadade01e805b01 Mar 14 09:30:47 crc kubenswrapper[4843]: I0314 09:30:47.349488 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c4c6c00-6874-4581-8123-8f0e506626a5" path="/var/lib/kubelet/pods/0c4c6c00-6874-4581-8123-8f0e506626a5/volumes" Mar 14 09:30:47 crc kubenswrapper[4843]: I0314 09:30:47.922508 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerStarted","Data":"076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986"} Mar 14 09:30:47 crc kubenswrapper[4843]: I0314 09:30:47.922780 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerStarted","Data":"53c47655c854f932df1376117113e85cb1b632b0878a0516fcadade01e805b01"} Mar 14 09:30:48 crc kubenswrapper[4843]: I0314 09:30:48.942753 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerStarted","Data":"38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9"} Mar 14 09:30:48 crc kubenswrapper[4843]: I0314 09:30:48.943076 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerStarted","Data":"a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0"} Mar 14 09:30:50 crc kubenswrapper[4843]: I0314 09:30:50.958176 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerStarted","Data":"72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec"} Mar 14 09:30:50 crc kubenswrapper[4843]: I0314 09:30:50.958304 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:30:50 crc kubenswrapper[4843]: I0314 09:30:50.984306 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.671499135 podStartE2EDuration="4.984285404s" podCreationTimestamp="2026-03-14 09:30:46 +0000 UTC" firstStartedPulling="2026-03-14 09:30:47.010291909 +0000 UTC m=+1214.322903037" lastFinishedPulling="2026-03-14 09:30:50.323078178 +0000 UTC m=+1217.635689306" observedRunningTime="2026-03-14 09:30:50.97815694 +0000 UTC m=+1218.290768078" watchObservedRunningTime="2026-03-14 09:30:50.984285404 +0000 UTC m=+1218.296896532" Mar 14 09:31:05 crc kubenswrapper[4843]: I0314 09:31:05.663878 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.443191 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.450561 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.452752 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"openstack-config-secret" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.452871 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"openstackclient-openstackclient-dockercfg-kdwm5" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.453362 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"openstack-config" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.454704 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.586972 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bknlk\" (UniqueName: \"kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.587394 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.587517 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.587730 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config-secret\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.598799 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:10 crc kubenswrapper[4843]: E0314 09:31:10.599517 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-bknlk openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="watcher-kuttl-default/openstackclient" podUID="313ea2a3-550d-4b97-abde-d64c14b822f4" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.607171 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.651356 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.652856 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.661962 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.689039 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.689105 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.689137 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config-secret\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.689173 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bknlk\" (UniqueName: \"kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.690533 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: E0314 09:31:10.695590 4843 projected.go:194] Error preparing data for projected volume kube-api-access-bknlk for pod watcher-kuttl-default/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (313ea2a3-550d-4b97-abde-d64c14b822f4) does not match the UID in record. The object might have been deleted and then recreated Mar 14 09:31:10 crc kubenswrapper[4843]: E0314 09:31:10.695669 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk podName:313ea2a3-550d-4b97-abde-d64c14b822f4 nodeName:}" failed. No retries permitted until 2026-03-14 09:31:11.195647138 +0000 UTC m=+1238.508258266 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bknlk" (UniqueName: "kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk") pod "openstackclient" (UID: "313ea2a3-550d-4b97-abde-d64c14b822f4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (313ea2a3-550d-4b97-abde-d64c14b822f4) does not match the UID in record. The object might have been deleted and then recreated Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.696079 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.709179 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config-secret\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.793436 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c3ef3e95-e2e0-456f-9262-2ae37fb42668-openstack-config-secret\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.793772 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ef3e95-e2e0-456f-9262-2ae37fb42668-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.793866 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6cp7\" (UniqueName: \"kubernetes.io/projected/c3ef3e95-e2e0-456f-9262-2ae37fb42668-kube-api-access-k6cp7\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.793992 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c3ef3e95-e2e0-456f-9262-2ae37fb42668-openstack-config\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.895170 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c3ef3e95-e2e0-456f-9262-2ae37fb42668-openstack-config\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.895294 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c3ef3e95-e2e0-456f-9262-2ae37fb42668-openstack-config-secret\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.895337 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ef3e95-e2e0-456f-9262-2ae37fb42668-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.895358 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6cp7\" (UniqueName: \"kubernetes.io/projected/c3ef3e95-e2e0-456f-9262-2ae37fb42668-kube-api-access-k6cp7\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.896248 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c3ef3e95-e2e0-456f-9262-2ae37fb42668-openstack-config\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.902914 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c3ef3e95-e2e0-456f-9262-2ae37fb42668-openstack-config-secret\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.910859 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6cp7\" (UniqueName: \"kubernetes.io/projected/c3ef3e95-e2e0-456f-9262-2ae37fb42668-kube-api-access-k6cp7\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.911369 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ef3e95-e2e0-456f-9262-2ae37fb42668-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c3ef3e95-e2e0-456f-9262-2ae37fb42668\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:10 crc kubenswrapper[4843]: I0314 09:31:10.967581 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.201501 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bknlk\" (UniqueName: \"kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk\") pod \"openstackclient\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:11 crc kubenswrapper[4843]: E0314 09:31:11.203950 4843 projected.go:194] Error preparing data for projected volume kube-api-access-bknlk for pod watcher-kuttl-default/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (313ea2a3-550d-4b97-abde-d64c14b822f4) does not match the UID in record. The object might have been deleted and then recreated Mar 14 09:31:11 crc kubenswrapper[4843]: E0314 09:31:11.204025 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk podName:313ea2a3-550d-4b97-abde-d64c14b822f4 nodeName:}" failed. No retries permitted until 2026-03-14 09:31:12.204005998 +0000 UTC m=+1239.516617126 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-bknlk" (UniqueName: "kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk") pod "openstackclient" (UID: "313ea2a3-550d-4b97-abde-d64c14b822f4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (313ea2a3-550d-4b97-abde-d64c14b822f4) does not match the UID in record. The object might have been deleted and then recreated Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.272330 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/openstackclient"] Mar 14 09:31:11 crc kubenswrapper[4843]: W0314 09:31:11.279830 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3ef3e95_e2e0_456f_9262_2ae37fb42668.slice/crio-989af54e3f4661c115da352b31e111186666a916713196b56c68d6ad4522ccf4 WatchSource:0}: Error finding container 989af54e3f4661c115da352b31e111186666a916713196b56c68d6ad4522ccf4: Status 404 returned error can't find the container with id 989af54e3f4661c115da352b31e111186666a916713196b56c68d6ad4522ccf4 Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.299082 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.299866 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstackclient" event={"ID":"c3ef3e95-e2e0-456f-9262-2ae37fb42668","Type":"ContainerStarted","Data":"989af54e3f4661c115da352b31e111186666a916713196b56c68d6ad4522ccf4"} Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.303488 4843 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="watcher-kuttl-default/openstackclient" oldPodUID="313ea2a3-550d-4b97-abde-d64c14b822f4" podUID="c3ef3e95-e2e0-456f-9262-2ae37fb42668" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.309740 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.404521 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-combined-ca-bundle\") pod \"313ea2a3-550d-4b97-abde-d64c14b822f4\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.404707 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config-secret\") pod \"313ea2a3-550d-4b97-abde-d64c14b822f4\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.404794 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config\") pod \"313ea2a3-550d-4b97-abde-d64c14b822f4\" (UID: \"313ea2a3-550d-4b97-abde-d64c14b822f4\") " Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.405238 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bknlk\" (UniqueName: \"kubernetes.io/projected/313ea2a3-550d-4b97-abde-d64c14b822f4-kube-api-access-bknlk\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.405809 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "313ea2a3-550d-4b97-abde-d64c14b822f4" (UID: "313ea2a3-550d-4b97-abde-d64c14b822f4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.409079 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "313ea2a3-550d-4b97-abde-d64c14b822f4" (UID: "313ea2a3-550d-4b97-abde-d64c14b822f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.409244 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "313ea2a3-550d-4b97-abde-d64c14b822f4" (UID: "313ea2a3-550d-4b97-abde-d64c14b822f4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.510300 4843 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.510344 4843 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/313ea2a3-550d-4b97-abde-d64c14b822f4-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:11 crc kubenswrapper[4843]: I0314 09:31:11.510354 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313ea2a3-550d-4b97-abde-d64c14b822f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:12 crc kubenswrapper[4843]: I0314 09:31:12.307795 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/openstackclient" Mar 14 09:31:12 crc kubenswrapper[4843]: I0314 09:31:12.327920 4843 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="watcher-kuttl-default/openstackclient" oldPodUID="313ea2a3-550d-4b97-abde-d64c14b822f4" podUID="c3ef3e95-e2e0-456f-9262-2ae37fb42668" Mar 14 09:31:13 crc kubenswrapper[4843]: I0314 09:31:13.350008 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="313ea2a3-550d-4b97-abde-d64c14b822f4" path="/var/lib/kubelet/pods/313ea2a3-550d-4b97-abde-d64c14b822f4/volumes" Mar 14 09:31:16 crc kubenswrapper[4843]: I0314 09:31:16.582113 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:19 crc kubenswrapper[4843]: I0314 09:31:19.687393 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:31:19 crc kubenswrapper[4843]: I0314 09:31:19.689769 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/kube-state-metrics-0" podUID="1c3ac2af-99e9-46e3-a447-5dcef51487ef" containerName="kube-state-metrics" containerID="cri-o://1d8d4d630e57b45c6c556e6b8945aa8ec049c786c5df18dac0aa554416c5882d" gracePeriod=30 Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.377196 4843 generic.go:334] "Generic (PLEG): container finished" podID="1c3ac2af-99e9-46e3-a447-5dcef51487ef" containerID="1d8d4d630e57b45c6c556e6b8945aa8ec049c786c5df18dac0aa554416c5882d" exitCode=2 Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.377288 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"1c3ac2af-99e9-46e3-a447-5dcef51487ef","Type":"ContainerDied","Data":"1d8d4d630e57b45c6c556e6b8945aa8ec049c786c5df18dac0aa554416c5882d"} Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.714120 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.714459 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-central-agent" containerID="cri-o://076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986" gracePeriod=30 Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.714586 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-notification-agent" containerID="cri-o://a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0" gracePeriod=30 Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.714559 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="sg-core" containerID="cri-o://38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9" gracePeriod=30 Mar 14 09:31:20 crc kubenswrapper[4843]: I0314 09:31:20.714712 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="proxy-httpd" containerID="cri-o://72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec" gracePeriod=30 Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.003051 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.076493 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whs92\" (UniqueName: \"kubernetes.io/projected/1c3ac2af-99e9-46e3-a447-5dcef51487ef-kube-api-access-whs92\") pod \"1c3ac2af-99e9-46e3-a447-5dcef51487ef\" (UID: \"1c3ac2af-99e9-46e3-a447-5dcef51487ef\") " Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.098869 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3ac2af-99e9-46e3-a447-5dcef51487ef-kube-api-access-whs92" (OuterVolumeSpecName: "kube-api-access-whs92") pod "1c3ac2af-99e9-46e3-a447-5dcef51487ef" (UID: "1c3ac2af-99e9-46e3-a447-5dcef51487ef"). InnerVolumeSpecName "kube-api-access-whs92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.179452 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whs92\" (UniqueName: \"kubernetes.io/projected/1c3ac2af-99e9-46e3-a447-5dcef51487ef-kube-api-access-whs92\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.389878 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/openstackclient" event={"ID":"c3ef3e95-e2e0-456f-9262-2ae37fb42668","Type":"ContainerStarted","Data":"ffb464ed35a1a591f6c0d4356df7a819487a09e8932e48e64534b1773eb2909e"} Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.392928 4843 generic.go:334] "Generic (PLEG): container finished" podID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerID="72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec" exitCode=0 Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.392966 4843 generic.go:334] "Generic (PLEG): container finished" podID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerID="38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9" exitCode=2 Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.392980 4843 generic.go:334] "Generic (PLEG): container finished" podID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerID="076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986" exitCode=0 Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.393031 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerDied","Data":"72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec"} Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.393062 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerDied","Data":"38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9"} Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.393076 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerDied","Data":"076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986"} Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.395067 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"1c3ac2af-99e9-46e3-a447-5dcef51487ef","Type":"ContainerDied","Data":"d458a056eebdc9a487d83ee2f986323f545797fdbbac141aafdd15369a9ba9a2"} Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.395117 4843 scope.go:117] "RemoveContainer" containerID="1d8d4d630e57b45c6c556e6b8945aa8ec049c786c5df18dac0aa554416c5882d" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.395223 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.410970 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/openstackclient" podStartSLOduration=1.934434085 podStartE2EDuration="11.410955647s" podCreationTimestamp="2026-03-14 09:31:10 +0000 UTC" firstStartedPulling="2026-03-14 09:31:11.28237668 +0000 UTC m=+1238.594987808" lastFinishedPulling="2026-03-14 09:31:20.758898242 +0000 UTC m=+1248.071509370" observedRunningTime="2026-03-14 09:31:21.409763929 +0000 UTC m=+1248.722375057" watchObservedRunningTime="2026-03-14 09:31:21.410955647 +0000 UTC m=+1248.723566775" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.443891 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.459615 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.467052 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:31:21 crc kubenswrapper[4843]: E0314 09:31:21.467473 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3ac2af-99e9-46e3-a447-5dcef51487ef" containerName="kube-state-metrics" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.467495 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3ac2af-99e9-46e3-a447-5dcef51487ef" containerName="kube-state-metrics" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.467666 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3ac2af-99e9-46e3-a447-5dcef51487ef" containerName="kube-state-metrics" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.468246 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.470238 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"kube-state-metrics-tls-config" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.470346 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-kube-state-metrics-svc" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.486383 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.585267 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.585429 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.585465 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.585544 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82w49\" (UniqueName: \"kubernetes.io/projected/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-api-access-82w49\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.686713 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.687031 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.687058 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.687125 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82w49\" (UniqueName: \"kubernetes.io/projected/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-api-access-82w49\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.697152 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.697257 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.697554 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.706777 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82w49\" (UniqueName: \"kubernetes.io/projected/4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced-kube-api-access-82w49\") pod \"kube-state-metrics-0\" (UID: \"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced\") " pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:21 crc kubenswrapper[4843]: I0314 09:31:21.786202 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:22 crc kubenswrapper[4843]: I0314 09:31:22.265495 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/kube-state-metrics-0"] Mar 14 09:31:22 crc kubenswrapper[4843]: I0314 09:31:22.403436 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced","Type":"ContainerStarted","Data":"a2675bc3c5a0fa1a3628e56dbb58785f908ef421522ee89a6c8e041f6480c583"} Mar 14 09:31:23 crc kubenswrapper[4843]: I0314 09:31:23.363508 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3ac2af-99e9-46e3-a447-5dcef51487ef" path="/var/lib/kubelet/pods/1c3ac2af-99e9-46e3-a447-5dcef51487ef/volumes" Mar 14 09:31:23 crc kubenswrapper[4843]: I0314 09:31:23.413993 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/kube-state-metrics-0" event={"ID":"4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced","Type":"ContainerStarted","Data":"a74d842e1a48cad41cde1fc0a204ca8c14f4f69a906a40786768b7665e82ec23"} Mar 14 09:31:23 crc kubenswrapper[4843]: I0314 09:31:23.414140 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:23 crc kubenswrapper[4843]: I0314 09:31:23.433939 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/kube-state-metrics-0" podStartSLOduration=2.042162149 podStartE2EDuration="2.433922389s" podCreationTimestamp="2026-03-14 09:31:21 +0000 UTC" firstStartedPulling="2026-03-14 09:31:22.272642513 +0000 UTC m=+1249.585253641" lastFinishedPulling="2026-03-14 09:31:22.664402743 +0000 UTC m=+1249.977013881" observedRunningTime="2026-03-14 09:31:23.428155502 +0000 UTC m=+1250.740766630" watchObservedRunningTime="2026-03-14 09:31:23.433922389 +0000 UTC m=+1250.746533517" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.243793 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.325827 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-combined-ca-bundle\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.325912 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl5wx\" (UniqueName: \"kubernetes.io/projected/008369ae-6b4f-4dbf-ab72-d279508616b7-kube-api-access-zl5wx\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.325970 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-config-data\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.326017 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-log-httpd\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.326051 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-run-httpd\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.326097 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-scripts\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.326153 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-sg-core-conf-yaml\") pod \"008369ae-6b4f-4dbf-ab72-d279508616b7\" (UID: \"008369ae-6b4f-4dbf-ab72-d279508616b7\") " Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.327114 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.327328 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.331525 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-scripts" (OuterVolumeSpecName: "scripts") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.331970 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/008369ae-6b4f-4dbf-ab72-d279508616b7-kube-api-access-zl5wx" (OuterVolumeSpecName: "kube-api-access-zl5wx") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "kube-api-access-zl5wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.351969 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.387492 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.425004 4843 generic.go:334] "Generic (PLEG): container finished" podID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerID="a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0" exitCode=0 Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.425067 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerDied","Data":"a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0"} Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.425134 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"008369ae-6b4f-4dbf-ab72-d279508616b7","Type":"ContainerDied","Data":"53c47655c854f932df1376117113e85cb1b632b0878a0516fcadade01e805b01"} Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.425158 4843 scope.go:117] "RemoveContainer" containerID="72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.425096 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.427404 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.427431 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/008369ae-6b4f-4dbf-ab72-d279508616b7-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.427442 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.427451 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.427461 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.427469 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl5wx\" (UniqueName: \"kubernetes.io/projected/008369ae-6b4f-4dbf-ab72-d279508616b7-kube-api-access-zl5wx\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.430992 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-config-data" (OuterVolumeSpecName: "config-data") pod "008369ae-6b4f-4dbf-ab72-d279508616b7" (UID: "008369ae-6b4f-4dbf-ab72-d279508616b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.445605 4843 scope.go:117] "RemoveContainer" containerID="38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.474536 4843 scope.go:117] "RemoveContainer" containerID="a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.492394 4843 scope.go:117] "RemoveContainer" containerID="076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.522106 4843 scope.go:117] "RemoveContainer" containerID="72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.522631 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec\": container with ID starting with 72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec not found: ID does not exist" containerID="72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.522674 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec"} err="failed to get container status \"72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec\": rpc error: code = NotFound desc = could not find container \"72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec\": container with ID starting with 72475f80550164814628dffb46e1ebd1399991f5fcdcc182194ab4872034c4ec not found: ID does not exist" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.522703 4843 scope.go:117] "RemoveContainer" containerID="38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.523132 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9\": container with ID starting with 38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9 not found: ID does not exist" containerID="38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.523165 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9"} err="failed to get container status \"38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9\": rpc error: code = NotFound desc = could not find container \"38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9\": container with ID starting with 38620a23f9e9984344b309e2e6c6f8fc10ed76bde111720fee5c57ded6c012d9 not found: ID does not exist" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.523187 4843 scope.go:117] "RemoveContainer" containerID="a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.523436 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0\": container with ID starting with a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0 not found: ID does not exist" containerID="a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.523456 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0"} err="failed to get container status \"a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0\": rpc error: code = NotFound desc = could not find container \"a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0\": container with ID starting with a7eacbcea8799457e71532878103a50155f0e38b790987926107e5cef50f5ed0 not found: ID does not exist" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.523469 4843 scope.go:117] "RemoveContainer" containerID="076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.523731 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986\": container with ID starting with 076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986 not found: ID does not exist" containerID="076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.523750 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986"} err="failed to get container status \"076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986\": rpc error: code = NotFound desc = could not find container \"076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986\": container with ID starting with 076e60288a18ca58499c059757f156dc683c91524a0979ddf78933ed1d4ea986 not found: ID does not exist" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.528855 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008369ae-6b4f-4dbf-ab72-d279508616b7-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.761972 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.769005 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.780565 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.780893 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="proxy-httpd" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.780914 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="proxy-httpd" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.780933 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="sg-core" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.780941 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="sg-core" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.780959 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-notification-agent" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.780965 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-notification-agent" Mar 14 09:31:24 crc kubenswrapper[4843]: E0314 09:31:24.780982 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-central-agent" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.780988 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-central-agent" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.781119 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-notification-agent" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.781135 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="ceilometer-central-agent" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.781144 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="sg-core" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.781156 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" containerName="proxy-httpd" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.782502 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.785063 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.785457 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.785911 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.800041 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934332 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-config-data\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934398 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934441 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934473 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-log-httpd\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934502 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-run-httpd\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934572 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-scripts\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934620 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:24 crc kubenswrapper[4843]: I0314 09:31:24.934649 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztvxm\" (UniqueName: \"kubernetes.io/projected/2267c349-e72b-4476-bc83-0c80ccafb1fc-kube-api-access-ztvxm\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.035847 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-config-data\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.036647 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.036761 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.036813 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-log-httpd\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.036878 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-run-httpd\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.037348 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-log-httpd\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.037445 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-run-httpd\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.036936 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-scripts\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.037514 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.037535 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztvxm\" (UniqueName: \"kubernetes.io/projected/2267c349-e72b-4476-bc83-0c80ccafb1fc-kube-api-access-ztvxm\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.040742 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-scripts\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.041627 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-config-data\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.043636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.054181 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.054886 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.061881 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztvxm\" (UniqueName: \"kubernetes.io/projected/2267c349-e72b-4476-bc83-0c80ccafb1fc-kube-api-access-ztvxm\") pod \"ceilometer-0\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.097965 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.348966 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="008369ae-6b4f-4dbf-ab72-d279508616b7" path="/var/lib/kubelet/pods/008369ae-6b4f-4dbf-ab72-d279508616b7/volumes" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.538351 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:31:25 crc kubenswrapper[4843]: W0314 09:31:25.540233 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2267c349_e72b_4476_bc83_0c80ccafb1fc.slice/crio-a25a9a953a4c4a470fe10af42499c912391642050b786f0fc622f71ea8ebefc2 WatchSource:0}: Error finding container a25a9a953a4c4a470fe10af42499c912391642050b786f0fc622f71ea8ebefc2: Status 404 returned error can't find the container with id a25a9a953a4c4a470fe10af42499c912391642050b786f0fc622f71ea8ebefc2 Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.828289 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-2pj9j"] Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.829982 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.841125 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-2pj9j"] Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.919363 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-6f63-account-create-update-dx22q"] Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.920622 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.925557 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.929063 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-6f63-account-create-update-dx22q"] Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.952418 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-operator-scripts\") pod \"watcher-db-create-2pj9j\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:25 crc kubenswrapper[4843]: I0314 09:31:25.952463 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqf52\" (UniqueName: \"kubernetes.io/projected/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-kube-api-access-lqf52\") pod \"watcher-db-create-2pj9j\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.054032 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9xx\" (UniqueName: \"kubernetes.io/projected/5378344b-d874-4589-aab2-78186101e8d0-kube-api-access-4w9xx\") pod \"watcher-6f63-account-create-update-dx22q\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.054321 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-operator-scripts\") pod \"watcher-db-create-2pj9j\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.054421 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5378344b-d874-4589-aab2-78186101e8d0-operator-scripts\") pod \"watcher-6f63-account-create-update-dx22q\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.054451 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqf52\" (UniqueName: \"kubernetes.io/projected/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-kube-api-access-lqf52\") pod \"watcher-db-create-2pj9j\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.055087 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-operator-scripts\") pod \"watcher-db-create-2pj9j\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.076989 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqf52\" (UniqueName: \"kubernetes.io/projected/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-kube-api-access-lqf52\") pod \"watcher-db-create-2pj9j\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.153459 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.156039 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9xx\" (UniqueName: \"kubernetes.io/projected/5378344b-d874-4589-aab2-78186101e8d0-kube-api-access-4w9xx\") pod \"watcher-6f63-account-create-update-dx22q\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.156175 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5378344b-d874-4589-aab2-78186101e8d0-operator-scripts\") pod \"watcher-6f63-account-create-update-dx22q\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.157043 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5378344b-d874-4589-aab2-78186101e8d0-operator-scripts\") pod \"watcher-6f63-account-create-update-dx22q\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.190147 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9xx\" (UniqueName: \"kubernetes.io/projected/5378344b-d874-4589-aab2-78186101e8d0-kube-api-access-4w9xx\") pod \"watcher-6f63-account-create-update-dx22q\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.316461 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.478899 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerStarted","Data":"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f"} Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.478981 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerStarted","Data":"a25a9a953a4c4a470fe10af42499c912391642050b786f0fc622f71ea8ebefc2"} Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.534257 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-2pj9j"] Mar 14 09:31:26 crc kubenswrapper[4843]: I0314 09:31:26.750346 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-6f63-account-create-update-dx22q"] Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.487737 4843 generic.go:334] "Generic (PLEG): container finished" podID="4765db79-5f59-4bdf-a86e-4f69fa09b1c0" containerID="903d9897aa84e59904dd95bc27b534b765bf5fe3a87107d2dbc1516d8b8cbdd7" exitCode=0 Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.487789 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-2pj9j" event={"ID":"4765db79-5f59-4bdf-a86e-4f69fa09b1c0","Type":"ContainerDied","Data":"903d9897aa84e59904dd95bc27b534b765bf5fe3a87107d2dbc1516d8b8cbdd7"} Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.487862 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-2pj9j" event={"ID":"4765db79-5f59-4bdf-a86e-4f69fa09b1c0","Type":"ContainerStarted","Data":"90faa4027d4bf5d144206e4d9e275af5fc4275856b093fd1a55b058749a2c975"} Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.489894 4843 generic.go:334] "Generic (PLEG): container finished" podID="5378344b-d874-4589-aab2-78186101e8d0" containerID="b08e7f51d48295ca3bcd9303e4b7ecbe20e705e3bc7ea69f20fd2986cd82addf" exitCode=0 Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.489970 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" event={"ID":"5378344b-d874-4589-aab2-78186101e8d0","Type":"ContainerDied","Data":"b08e7f51d48295ca3bcd9303e4b7ecbe20e705e3bc7ea69f20fd2986cd82addf"} Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.489999 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" event={"ID":"5378344b-d874-4589-aab2-78186101e8d0","Type":"ContainerStarted","Data":"607e185b777df2ee8aafaf4319c899811197a3707c6a486192675e28a023afb3"} Mar 14 09:31:27 crc kubenswrapper[4843]: I0314 09:31:27.491958 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerStarted","Data":"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509"} Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.502221 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerStarted","Data":"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e"} Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.806996 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.906906 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w9xx\" (UniqueName: \"kubernetes.io/projected/5378344b-d874-4589-aab2-78186101e8d0-kube-api-access-4w9xx\") pod \"5378344b-d874-4589-aab2-78186101e8d0\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.907096 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5378344b-d874-4589-aab2-78186101e8d0-operator-scripts\") pod \"5378344b-d874-4589-aab2-78186101e8d0\" (UID: \"5378344b-d874-4589-aab2-78186101e8d0\") " Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.907882 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5378344b-d874-4589-aab2-78186101e8d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5378344b-d874-4589-aab2-78186101e8d0" (UID: "5378344b-d874-4589-aab2-78186101e8d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.913977 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5378344b-d874-4589-aab2-78186101e8d0-kube-api-access-4w9xx" (OuterVolumeSpecName: "kube-api-access-4w9xx") pod "5378344b-d874-4589-aab2-78186101e8d0" (UID: "5378344b-d874-4589-aab2-78186101e8d0"). InnerVolumeSpecName "kube-api-access-4w9xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:31:28 crc kubenswrapper[4843]: I0314 09:31:28.948802 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.008814 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqf52\" (UniqueName: \"kubernetes.io/projected/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-kube-api-access-lqf52\") pod \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.008964 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-operator-scripts\") pod \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\" (UID: \"4765db79-5f59-4bdf-a86e-4f69fa09b1c0\") " Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.009309 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w9xx\" (UniqueName: \"kubernetes.io/projected/5378344b-d874-4589-aab2-78186101e8d0-kube-api-access-4w9xx\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.009324 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5378344b-d874-4589-aab2-78186101e8d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.009362 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4765db79-5f59-4bdf-a86e-4f69fa09b1c0" (UID: "4765db79-5f59-4bdf-a86e-4f69fa09b1c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.012481 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-kube-api-access-lqf52" (OuterVolumeSpecName: "kube-api-access-lqf52") pod "4765db79-5f59-4bdf-a86e-4f69fa09b1c0" (UID: "4765db79-5f59-4bdf-a86e-4f69fa09b1c0"). InnerVolumeSpecName "kube-api-access-lqf52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.110954 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqf52\" (UniqueName: \"kubernetes.io/projected/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-kube-api-access-lqf52\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.110993 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4765db79-5f59-4bdf-a86e-4f69fa09b1c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.511165 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-2pj9j" event={"ID":"4765db79-5f59-4bdf-a86e-4f69fa09b1c0","Type":"ContainerDied","Data":"90faa4027d4bf5d144206e4d9e275af5fc4275856b093fd1a55b058749a2c975"} Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.511451 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90faa4027d4bf5d144206e4d9e275af5fc4275856b093fd1a55b058749a2c975" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.511181 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-2pj9j" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.514306 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" event={"ID":"5378344b-d874-4589-aab2-78186101e8d0","Type":"ContainerDied","Data":"607e185b777df2ee8aafaf4319c899811197a3707c6a486192675e28a023afb3"} Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.514346 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-6f63-account-create-update-dx22q" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.514355 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="607e185b777df2ee8aafaf4319c899811197a3707c6a486192675e28a023afb3" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.516747 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerStarted","Data":"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7"} Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.517176 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:31:29 crc kubenswrapper[4843]: I0314 09:31:29.548994 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.9651105169999998 podStartE2EDuration="5.548973991s" podCreationTimestamp="2026-03-14 09:31:24 +0000 UTC" firstStartedPulling="2026-03-14 09:31:25.543233282 +0000 UTC m=+1252.855844410" lastFinishedPulling="2026-03-14 09:31:29.127096756 +0000 UTC m=+1256.439707884" observedRunningTime="2026-03-14 09:31:29.544830013 +0000 UTC m=+1256.857441161" watchObservedRunningTime="2026-03-14 09:31:29.548973991 +0000 UTC m=+1256.861585139" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.254321 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sc292"] Mar 14 09:31:31 crc kubenswrapper[4843]: E0314 09:31:31.255113 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5378344b-d874-4589-aab2-78186101e8d0" containerName="mariadb-account-create-update" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.255138 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5378344b-d874-4589-aab2-78186101e8d0" containerName="mariadb-account-create-update" Mar 14 09:31:31 crc kubenswrapper[4843]: E0314 09:31:31.255153 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4765db79-5f59-4bdf-a86e-4f69fa09b1c0" containerName="mariadb-database-create" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.255159 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4765db79-5f59-4bdf-a86e-4f69fa09b1c0" containerName="mariadb-database-create" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.255386 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4765db79-5f59-4bdf-a86e-4f69fa09b1c0" containerName="mariadb-database-create" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.255416 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5378344b-d874-4589-aab2-78186101e8d0" containerName="mariadb-account-create-update" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.256320 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.258085 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-vpfpc" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.259636 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.262883 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sc292"] Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.345371 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.345906 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-db-sync-config-data\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.346016 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-config-data\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.346210 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk86r\" (UniqueName: \"kubernetes.io/projected/a3bee036-655a-484c-b818-12bbdb0ddfe5-kube-api-access-fk86r\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.448238 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk86r\" (UniqueName: \"kubernetes.io/projected/a3bee036-655a-484c-b818-12bbdb0ddfe5-kube-api-access-fk86r\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.448391 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.448471 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-db-sync-config-data\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.448513 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-config-data\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.454103 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.454356 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-config-data\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.454670 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-db-sync-config-data\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.467993 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk86r\" (UniqueName: \"kubernetes.io/projected/a3bee036-655a-484c-b818-12bbdb0ddfe5-kube-api-access-fk86r\") pod \"watcher-kuttl-db-sync-sc292\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.584658 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:31:31 crc kubenswrapper[4843]: I0314 09:31:31.802921 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/kube-state-metrics-0" Mar 14 09:31:32 crc kubenswrapper[4843]: I0314 09:31:32.055458 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sc292"] Mar 14 09:31:32 crc kubenswrapper[4843]: I0314 09:31:32.550890 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" event={"ID":"a3bee036-655a-484c-b818-12bbdb0ddfe5","Type":"ContainerStarted","Data":"48cc0049c35be2765bcf7ec91f09346e541263a126787ee124f83a74945e34d8"} Mar 14 09:31:48 crc kubenswrapper[4843]: E0314 09:31:48.280188 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.243:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Mar 14 09:31:48 crc kubenswrapper[4843]: E0314 09:31:48.281401 4843 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.243:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Mar 14 09:31:48 crc kubenswrapper[4843]: E0314 09:31:48.281556 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-kuttl-db-sync,Image:38.102.83.243:5001/podified-master-centos10/openstack-watcher-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/watcher/watcher.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fk86r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-kuttl-db-sync-sc292_watcher-kuttl-default(a3bee036-655a-484c-b818-12bbdb0ddfe5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:31:48 crc kubenswrapper[4843]: E0314 09:31:48.282833 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-kuttl-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" podUID="a3bee036-655a-484c-b818-12bbdb0ddfe5" Mar 14 09:31:48 crc kubenswrapper[4843]: E0314 09:31:48.722663 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-kuttl-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.243:5001/podified-master-centos10/openstack-watcher-api:watcher_latest\\\"\"" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" podUID="a3bee036-655a-484c-b818-12bbdb0ddfe5" Mar 14 09:31:55 crc kubenswrapper[4843]: I0314 09:31:55.270757 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.145739 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558012-nrxkk"] Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.147570 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.151105 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.151373 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.151512 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.155713 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558012-nrxkk"] Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.238157 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5l6\" (UniqueName: \"kubernetes.io/projected/74cd6ce3-6aba-48e9-a268-76deeffbadb5-kube-api-access-vl5l6\") pod \"auto-csr-approver-29558012-nrxkk\" (UID: \"74cd6ce3-6aba-48e9-a268-76deeffbadb5\") " pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.339640 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5l6\" (UniqueName: \"kubernetes.io/projected/74cd6ce3-6aba-48e9-a268-76deeffbadb5-kube-api-access-vl5l6\") pod \"auto-csr-approver-29558012-nrxkk\" (UID: \"74cd6ce3-6aba-48e9-a268-76deeffbadb5\") " pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.360941 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5l6\" (UniqueName: \"kubernetes.io/projected/74cd6ce3-6aba-48e9-a268-76deeffbadb5-kube-api-access-vl5l6\") pod \"auto-csr-approver-29558012-nrxkk\" (UID: \"74cd6ce3-6aba-48e9-a268-76deeffbadb5\") " pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.473815 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:00 crc kubenswrapper[4843]: I0314 09:32:00.909821 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558012-nrxkk"] Mar 14 09:32:01 crc kubenswrapper[4843]: I0314 09:32:01.818853 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" event={"ID":"74cd6ce3-6aba-48e9-a268-76deeffbadb5","Type":"ContainerStarted","Data":"0779b8516b3519321bf49cd6b82bc012e1eb3db9a373efe9320edf2e243eff7d"} Mar 14 09:32:02 crc kubenswrapper[4843]: I0314 09:32:02.828050 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" event={"ID":"74cd6ce3-6aba-48e9-a268-76deeffbadb5","Type":"ContainerStarted","Data":"627afe853906b53970a37d1adcd44701fb7f26abbf7086812f1278f110f77a7b"} Mar 14 09:32:02 crc kubenswrapper[4843]: I0314 09:32:02.829706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" event={"ID":"a3bee036-655a-484c-b818-12bbdb0ddfe5","Type":"ContainerStarted","Data":"1fc7936c99da80ed1c59476ba191691097198e4678919c894a43b157662fb41e"} Mar 14 09:32:02 crc kubenswrapper[4843]: I0314 09:32:02.841752 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" podStartSLOduration=1.358504483 podStartE2EDuration="2.84173256s" podCreationTimestamp="2026-03-14 09:32:00 +0000 UTC" firstStartedPulling="2026-03-14 09:32:00.919307867 +0000 UTC m=+1288.231918995" lastFinishedPulling="2026-03-14 09:32:02.402535944 +0000 UTC m=+1289.715147072" observedRunningTime="2026-03-14 09:32:02.84172703 +0000 UTC m=+1290.154338168" watchObservedRunningTime="2026-03-14 09:32:02.84173256 +0000 UTC m=+1290.154343688" Mar 14 09:32:02 crc kubenswrapper[4843]: I0314 09:32:02.863995 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" podStartSLOduration=1.476103317 podStartE2EDuration="31.863974798s" podCreationTimestamp="2026-03-14 09:31:31 +0000 UTC" firstStartedPulling="2026-03-14 09:31:32.057955422 +0000 UTC m=+1259.370566550" lastFinishedPulling="2026-03-14 09:32:02.445826903 +0000 UTC m=+1289.758438031" observedRunningTime="2026-03-14 09:32:02.859252417 +0000 UTC m=+1290.171863545" watchObservedRunningTime="2026-03-14 09:32:02.863974798 +0000 UTC m=+1290.176585926" Mar 14 09:32:03 crc kubenswrapper[4843]: I0314 09:32:03.838927 4843 generic.go:334] "Generic (PLEG): container finished" podID="74cd6ce3-6aba-48e9-a268-76deeffbadb5" containerID="627afe853906b53970a37d1adcd44701fb7f26abbf7086812f1278f110f77a7b" exitCode=0 Mar 14 09:32:03 crc kubenswrapper[4843]: I0314 09:32:03.838971 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" event={"ID":"74cd6ce3-6aba-48e9-a268-76deeffbadb5","Type":"ContainerDied","Data":"627afe853906b53970a37d1adcd44701fb7f26abbf7086812f1278f110f77a7b"} Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.187996 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.231921 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl5l6\" (UniqueName: \"kubernetes.io/projected/74cd6ce3-6aba-48e9-a268-76deeffbadb5-kube-api-access-vl5l6\") pod \"74cd6ce3-6aba-48e9-a268-76deeffbadb5\" (UID: \"74cd6ce3-6aba-48e9-a268-76deeffbadb5\") " Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.240026 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74cd6ce3-6aba-48e9-a268-76deeffbadb5-kube-api-access-vl5l6" (OuterVolumeSpecName: "kube-api-access-vl5l6") pod "74cd6ce3-6aba-48e9-a268-76deeffbadb5" (UID: "74cd6ce3-6aba-48e9-a268-76deeffbadb5"). InnerVolumeSpecName "kube-api-access-vl5l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.333791 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl5l6\" (UniqueName: \"kubernetes.io/projected/74cd6ce3-6aba-48e9-a268-76deeffbadb5-kube-api-access-vl5l6\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.857229 4843 generic.go:334] "Generic (PLEG): container finished" podID="a3bee036-655a-484c-b818-12bbdb0ddfe5" containerID="1fc7936c99da80ed1c59476ba191691097198e4678919c894a43b157662fb41e" exitCode=0 Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.857325 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" event={"ID":"a3bee036-655a-484c-b818-12bbdb0ddfe5","Type":"ContainerDied","Data":"1fc7936c99da80ed1c59476ba191691097198e4678919c894a43b157662fb41e"} Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.858965 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" event={"ID":"74cd6ce3-6aba-48e9-a268-76deeffbadb5","Type":"ContainerDied","Data":"0779b8516b3519321bf49cd6b82bc012e1eb3db9a373efe9320edf2e243eff7d"} Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.859001 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0779b8516b3519321bf49cd6b82bc012e1eb3db9a373efe9320edf2e243eff7d" Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.859063 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558012-nrxkk" Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.919836 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558006-9fnwg"] Mar 14 09:32:05 crc kubenswrapper[4843]: I0314 09:32:05.925180 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558006-9fnwg"] Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.121708 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.162119 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-combined-ca-bundle\") pod \"a3bee036-655a-484c-b818-12bbdb0ddfe5\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.162357 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk86r\" (UniqueName: \"kubernetes.io/projected/a3bee036-655a-484c-b818-12bbdb0ddfe5-kube-api-access-fk86r\") pod \"a3bee036-655a-484c-b818-12bbdb0ddfe5\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.162417 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-db-sync-config-data\") pod \"a3bee036-655a-484c-b818-12bbdb0ddfe5\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.162476 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-config-data\") pod \"a3bee036-655a-484c-b818-12bbdb0ddfe5\" (UID: \"a3bee036-655a-484c-b818-12bbdb0ddfe5\") " Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.167523 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a3bee036-655a-484c-b818-12bbdb0ddfe5" (UID: "a3bee036-655a-484c-b818-12bbdb0ddfe5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.180346 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3bee036-655a-484c-b818-12bbdb0ddfe5-kube-api-access-fk86r" (OuterVolumeSpecName: "kube-api-access-fk86r") pod "a3bee036-655a-484c-b818-12bbdb0ddfe5" (UID: "a3bee036-655a-484c-b818-12bbdb0ddfe5"). InnerVolumeSpecName "kube-api-access-fk86r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.194495 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3bee036-655a-484c-b818-12bbdb0ddfe5" (UID: "a3bee036-655a-484c-b818-12bbdb0ddfe5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.222599 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-config-data" (OuterVolumeSpecName: "config-data") pod "a3bee036-655a-484c-b818-12bbdb0ddfe5" (UID: "a3bee036-655a-484c-b818-12bbdb0ddfe5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.264545 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk86r\" (UniqueName: \"kubernetes.io/projected/a3bee036-655a-484c-b818-12bbdb0ddfe5-kube-api-access-fk86r\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.264808 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.264888 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.264967 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bee036-655a-484c-b818-12bbdb0ddfe5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.347907 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f044e544-ec72-408a-9848-b98f70c0669a" path="/var/lib/kubelet/pods/f044e544-ec72-408a-9848-b98f70c0669a/volumes" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.873684 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" event={"ID":"a3bee036-655a-484c-b818-12bbdb0ddfe5","Type":"ContainerDied","Data":"48cc0049c35be2765bcf7ec91f09346e541263a126787ee124f83a74945e34d8"} Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.873719 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48cc0049c35be2765bcf7ec91f09346e541263a126787ee124f83a74945e34d8" Mar 14 09:32:07 crc kubenswrapper[4843]: I0314 09:32:07.873756 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sc292" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.222828 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:32:08 crc kubenswrapper[4843]: E0314 09:32:08.223806 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bee036-655a-484c-b818-12bbdb0ddfe5" containerName="watcher-kuttl-db-sync" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.223884 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bee036-655a-484c-b818-12bbdb0ddfe5" containerName="watcher-kuttl-db-sync" Mar 14 09:32:08 crc kubenswrapper[4843]: E0314 09:32:08.223964 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cd6ce3-6aba-48e9-a268-76deeffbadb5" containerName="oc" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.224017 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cd6ce3-6aba-48e9-a268-76deeffbadb5" containerName="oc" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.224208 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="74cd6ce3-6aba-48e9-a268-76deeffbadb5" containerName="oc" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.224289 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3bee036-655a-484c-b818-12bbdb0ddfe5" containerName="watcher-kuttl-db-sync" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.225357 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.233015 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-vpfpc" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.233737 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.238234 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.282541 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbqsw\" (UniqueName: \"kubernetes.io/projected/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-kube-api-access-fbqsw\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.282633 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.282753 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-logs\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.282883 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.282916 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.333974 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.335360 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.344337 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.370758 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.377784 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.381290 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.383919 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.384409 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.384482 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.392584 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.398560 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.402426 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbqsw\" (UniqueName: \"kubernetes.io/projected/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-kube-api-access-fbqsw\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.402572 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.402630 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-logs\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.403083 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-logs\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.411304 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.435302 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbqsw\" (UniqueName: \"kubernetes.io/projected/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-kube-api-access-fbqsw\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.436050 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503657 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503724 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503771 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4b4q\" (UniqueName: \"kubernetes.io/projected/b7e25e6d-3253-4ac0-a420-3452f4f8275f-kube-api-access-p4b4q\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503795 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503842 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7e25e6d-3253-4ac0-a420-3452f4f8275f-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503863 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503884 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psf2n\" (UniqueName: \"kubernetes.io/projected/80eda2d0-78f4-4bac-b183-5ecf731131bb-kube-api-access-psf2n\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503902 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80eda2d0-78f4-4bac-b183-5ecf731131bb-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.503943 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.581923 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.604843 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80eda2d0-78f4-4bac-b183-5ecf731131bb-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.604924 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.604950 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.604998 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.605040 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4b4q\" (UniqueName: \"kubernetes.io/projected/b7e25e6d-3253-4ac0-a420-3452f4f8275f-kube-api-access-p4b4q\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.605339 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80eda2d0-78f4-4bac-b183-5ecf731131bb-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.605829 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.605862 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7e25e6d-3253-4ac0-a420-3452f4f8275f-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.605885 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.605905 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psf2n\" (UniqueName: \"kubernetes.io/projected/80eda2d0-78f4-4bac-b183-5ecf731131bb-kube-api-access-psf2n\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.606286 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7e25e6d-3253-4ac0-a420-3452f4f8275f-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.608430 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.611657 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.618117 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.618735 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.629173 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.635724 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psf2n\" (UniqueName: \"kubernetes.io/projected/80eda2d0-78f4-4bac-b183-5ecf731131bb-kube-api-access-psf2n\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.637014 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4b4q\" (UniqueName: \"kubernetes.io/projected/b7e25e6d-3253-4ac0-a420-3452f4f8275f-kube-api-access-p4b4q\") pod \"watcher-kuttl-applier-0\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.689687 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:08 crc kubenswrapper[4843]: I0314 09:32:08.795948 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:09 crc kubenswrapper[4843]: I0314 09:32:09.224348 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:32:09 crc kubenswrapper[4843]: I0314 09:32:09.233865 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:32:09 crc kubenswrapper[4843]: W0314 09:32:09.237247 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7e25e6d_3253_4ac0_a420_3452f4f8275f.slice/crio-f871e7e63af74bc13a158d279cefcd719c29af4053d04db452b6dc4967fd8e00 WatchSource:0}: Error finding container f871e7e63af74bc13a158d279cefcd719c29af4053d04db452b6dc4967fd8e00: Status 404 returned error can't find the container with id f871e7e63af74bc13a158d279cefcd719c29af4053d04db452b6dc4967fd8e00 Mar 14 09:32:09 crc kubenswrapper[4843]: I0314 09:32:09.328580 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:32:09 crc kubenswrapper[4843]: I0314 09:32:09.898005 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b7e25e6d-3253-4ac0-a420-3452f4f8275f","Type":"ContainerStarted","Data":"f871e7e63af74bc13a158d279cefcd719c29af4053d04db452b6dc4967fd8e00"} Mar 14 09:32:09 crc kubenswrapper[4843]: I0314 09:32:09.899629 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"80eda2d0-78f4-4bac-b183-5ecf731131bb","Type":"ContainerStarted","Data":"d43df2b06c99a71f87f6edce513b583fb32028f5b14cc17d60ea11cc40cb5428"} Mar 14 09:32:09 crc kubenswrapper[4843]: I0314 09:32:09.901437 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"33a65ce6-9c33-447c-b19a-17ab9bdadbd4","Type":"ContainerStarted","Data":"45c0c669520d9f636e1b1ec8a84d7b6489cf12a7e23f96393e0cf05ca2331444"} Mar 14 09:32:12 crc kubenswrapper[4843]: I0314 09:32:12.930712 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"33a65ce6-9c33-447c-b19a-17ab9bdadbd4","Type":"ContainerStarted","Data":"7768f365244cf747cf507e0a18759142f1e50773fa5d722a07446ad397f247b2"} Mar 14 09:32:13 crc kubenswrapper[4843]: I0314 09:32:13.941264 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"33a65ce6-9c33-447c-b19a-17ab9bdadbd4","Type":"ContainerStarted","Data":"f531fb376f64a529edacba1abe9989f5e77a7bf1713f5faf46b594293d49955d"} Mar 14 09:32:13 crc kubenswrapper[4843]: I0314 09:32:13.942198 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:13 crc kubenswrapper[4843]: I0314 09:32:13.960690 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=5.960667451 podStartE2EDuration="5.960667451s" podCreationTimestamp="2026-03-14 09:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:32:13.958458268 +0000 UTC m=+1301.271069396" watchObservedRunningTime="2026-03-14 09:32:13.960667451 +0000 UTC m=+1301.273278579" Mar 14 09:32:14 crc kubenswrapper[4843]: I0314 09:32:14.962689 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b7e25e6d-3253-4ac0-a420-3452f4f8275f","Type":"ContainerStarted","Data":"0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a"} Mar 14 09:32:14 crc kubenswrapper[4843]: I0314 09:32:14.965996 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"80eda2d0-78f4-4bac-b183-5ecf731131bb","Type":"ContainerStarted","Data":"6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45"} Mar 14 09:32:15 crc kubenswrapper[4843]: I0314 09:32:15.007729 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.3840572890000002 podStartE2EDuration="7.007708732s" podCreationTimestamp="2026-03-14 09:32:08 +0000 UTC" firstStartedPulling="2026-03-14 09:32:09.239020839 +0000 UTC m=+1296.551631977" lastFinishedPulling="2026-03-14 09:32:13.862672292 +0000 UTC m=+1301.175283420" observedRunningTime="2026-03-14 09:32:14.982069572 +0000 UTC m=+1302.294680690" watchObservedRunningTime="2026-03-14 09:32:15.007708732 +0000 UTC m=+1302.320319860" Mar 14 09:32:15 crc kubenswrapper[4843]: I0314 09:32:15.023498 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.556754034 podStartE2EDuration="7.023483267s" podCreationTimestamp="2026-03-14 09:32:08 +0000 UTC" firstStartedPulling="2026-03-14 09:32:09.3997747 +0000 UTC m=+1296.712385828" lastFinishedPulling="2026-03-14 09:32:13.866503933 +0000 UTC m=+1301.179115061" observedRunningTime="2026-03-14 09:32:15.014321578 +0000 UTC m=+1302.326932696" watchObservedRunningTime="2026-03-14 09:32:15.023483267 +0000 UTC m=+1302.336094395" Mar 14 09:32:15 crc kubenswrapper[4843]: I0314 09:32:15.387220 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:32:15 crc kubenswrapper[4843]: I0314 09:32:15.387324 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:32:16 crc kubenswrapper[4843]: I0314 09:32:16.268804 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.583141 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.583854 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.589964 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.691153 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.691225 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.719374 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.796550 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:18 crc kubenswrapper[4843]: I0314 09:32:18.821066 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:19 crc kubenswrapper[4843]: I0314 09:32:19.004085 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:19 crc kubenswrapper[4843]: I0314 09:32:19.008771 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:32:19 crc kubenswrapper[4843]: I0314 09:32:19.035588 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:32:19 crc kubenswrapper[4843]: I0314 09:32:19.054466 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.087177 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.087525 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-central-agent" containerID="cri-o://bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" gracePeriod=30 Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.087582 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="proxy-httpd" containerID="cri-o://3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" gracePeriod=30 Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.087613 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="sg-core" containerID="cri-o://303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" gracePeriod=30 Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.087662 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-notification-agent" containerID="cri-o://b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" gracePeriod=30 Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.874621 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970580 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-log-httpd\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970673 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-ceilometer-tls-certs\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970725 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-config-data\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970792 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-scripts\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970822 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-combined-ca-bundle\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970843 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-sg-core-conf-yaml\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970913 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztvxm\" (UniqueName: \"kubernetes.io/projected/2267c349-e72b-4476-bc83-0c80ccafb1fc-kube-api-access-ztvxm\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.970966 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-run-httpd\") pod \"2267c349-e72b-4476-bc83-0c80ccafb1fc\" (UID: \"2267c349-e72b-4476-bc83-0c80ccafb1fc\") " Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.971416 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.971660 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:32:21 crc kubenswrapper[4843]: I0314 09:32:21.978481 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-scripts" (OuterVolumeSpecName: "scripts") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.002853 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2267c349-e72b-4476-bc83-0c80ccafb1fc-kube-api-access-ztvxm" (OuterVolumeSpecName: "kube-api-access-ztvxm") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "kube-api-access-ztvxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.016785 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031099 4843 generic.go:334] "Generic (PLEG): container finished" podID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" exitCode=0 Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031127 4843 generic.go:334] "Generic (PLEG): container finished" podID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" exitCode=2 Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031137 4843 generic.go:334] "Generic (PLEG): container finished" podID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" exitCode=0 Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031148 4843 generic.go:334] "Generic (PLEG): container finished" podID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" exitCode=0 Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031143 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerDied","Data":"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7"} Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031187 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerDied","Data":"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e"} Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031202 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerDied","Data":"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509"} Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031213 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerDied","Data":"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f"} Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031225 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2267c349-e72b-4476-bc83-0c80ccafb1fc","Type":"ContainerDied","Data":"a25a9a953a4c4a470fe10af42499c912391642050b786f0fc622f71ea8ebefc2"} Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031249 4843 scope.go:117] "RemoveContainer" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.031749 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.051706 4843 scope.go:117] "RemoveContainer" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.052183 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072520 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072546 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072557 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztvxm\" (UniqueName: \"kubernetes.io/projected/2267c349-e72b-4476-bc83-0c80ccafb1fc-kube-api-access-ztvxm\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072566 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072574 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2267c349-e72b-4476-bc83-0c80ccafb1fc-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072583 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.072941 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.074382 4843 scope.go:117] "RemoveContainer" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.094848 4843 scope.go:117] "RemoveContainer" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.096221 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-config-data" (OuterVolumeSpecName: "config-data") pod "2267c349-e72b-4476-bc83-0c80ccafb1fc" (UID: "2267c349-e72b-4476-bc83-0c80ccafb1fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.117195 4843 scope.go:117] "RemoveContainer" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.117653 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": container with ID starting with 3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7 not found: ID does not exist" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.117694 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7"} err="failed to get container status \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": rpc error: code = NotFound desc = could not find container \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": container with ID starting with 3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.117718 4843 scope.go:117] "RemoveContainer" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.118507 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": container with ID starting with 303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e not found: ID does not exist" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.118537 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e"} err="failed to get container status \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": rpc error: code = NotFound desc = could not find container \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": container with ID starting with 303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.118563 4843 scope.go:117] "RemoveContainer" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.118823 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": container with ID starting with b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509 not found: ID does not exist" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.118842 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509"} err="failed to get container status \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": rpc error: code = NotFound desc = could not find container \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": container with ID starting with b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.118855 4843 scope.go:117] "RemoveContainer" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.119256 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": container with ID starting with bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f not found: ID does not exist" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.119486 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f"} err="failed to get container status \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": rpc error: code = NotFound desc = could not find container \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": container with ID starting with bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.119503 4843 scope.go:117] "RemoveContainer" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.119741 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7"} err="failed to get container status \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": rpc error: code = NotFound desc = could not find container \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": container with ID starting with 3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.119765 4843 scope.go:117] "RemoveContainer" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.119971 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e"} err="failed to get container status \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": rpc error: code = NotFound desc = could not find container \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": container with ID starting with 303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.119989 4843 scope.go:117] "RemoveContainer" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.120215 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509"} err="failed to get container status \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": rpc error: code = NotFound desc = could not find container \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": container with ID starting with b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.120241 4843 scope.go:117] "RemoveContainer" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.120596 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f"} err="failed to get container status \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": rpc error: code = NotFound desc = could not find container \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": container with ID starting with bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.120617 4843 scope.go:117] "RemoveContainer" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.121064 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7"} err="failed to get container status \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": rpc error: code = NotFound desc = could not find container \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": container with ID starting with 3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.121088 4843 scope.go:117] "RemoveContainer" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.121963 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e"} err="failed to get container status \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": rpc error: code = NotFound desc = could not find container \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": container with ID starting with 303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.122044 4843 scope.go:117] "RemoveContainer" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.122391 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509"} err="failed to get container status \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": rpc error: code = NotFound desc = could not find container \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": container with ID starting with b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.122440 4843 scope.go:117] "RemoveContainer" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.122707 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f"} err="failed to get container status \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": rpc error: code = NotFound desc = could not find container \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": container with ID starting with bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.122774 4843 scope.go:117] "RemoveContainer" containerID="3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123016 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7"} err="failed to get container status \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": rpc error: code = NotFound desc = could not find container \"3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7\": container with ID starting with 3f54baf4c64cfd8234a21599550eacb324b2a8163fd84e70bf419538a0d149f7 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123039 4843 scope.go:117] "RemoveContainer" containerID="303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123293 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e"} err="failed to get container status \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": rpc error: code = NotFound desc = could not find container \"303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e\": container with ID starting with 303de6589e12524e732179493a25229c5f17487ddda7e93fff1c72dc1cb3985e not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123357 4843 scope.go:117] "RemoveContainer" containerID="b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123661 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509"} err="failed to get container status \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": rpc error: code = NotFound desc = could not find container \"b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509\": container with ID starting with b5ec3cf91ba42617a1e5184701331c9c00b97f25cc010afc92df39c8a4904509 not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123726 4843 scope.go:117] "RemoveContainer" containerID="bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.123999 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f"} err="failed to get container status \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": rpc error: code = NotFound desc = could not find container \"bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f\": container with ID starting with bc3860669a3e68da66b2e1b4ae50a47ab5074339ce5ce7ea6fdf9cbbbe6f158f not found: ID does not exist" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.174506 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.174761 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2267c349-e72b-4476-bc83-0c80ccafb1fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.369167 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.377257 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.414140 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.414863 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="sg-core" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.414888 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="sg-core" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.414919 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-notification-agent" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.414930 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-notification-agent" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.414980 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-central-agent" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.414990 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-central-agent" Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.415006 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="proxy-httpd" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.415014 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="proxy-httpd" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.415467 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="proxy-httpd" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.415496 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="sg-core" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.415518 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-central-agent" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.415538 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" containerName="ceilometer-notification-agent" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.422693 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.427732 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.427960 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.428184 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.452708 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:32:22 crc kubenswrapper[4843]: E0314 09:32:22.467081 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2267c349_e72b_4476_bc83_0c80ccafb1fc.slice\": RecentStats: unable to find data in memory cache]" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581031 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdzzt\" (UniqueName: \"kubernetes.io/projected/18cb7825-ada4-4690-a848-9e8379bb7237-kube-api-access-rdzzt\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581108 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-run-httpd\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581141 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-config-data\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581439 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581603 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-log-httpd\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581661 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-scripts\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581812 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.581880 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.682970 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-log-httpd\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683039 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-scripts\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683086 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683123 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683172 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdzzt\" (UniqueName: \"kubernetes.io/projected/18cb7825-ada4-4690-a848-9e8379bb7237-kube-api-access-rdzzt\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683200 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-run-httpd\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683226 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-config-data\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683304 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683453 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-log-httpd\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.683720 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-run-httpd\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.686717 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.686772 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.686936 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.697973 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-scripts\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.700529 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-config-data\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.700773 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdzzt\" (UniqueName: \"kubernetes.io/projected/18cb7825-ada4-4690-a848-9e8379bb7237-kube-api-access-rdzzt\") pod \"ceilometer-0\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:22 crc kubenswrapper[4843]: I0314 09:32:22.761291 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:23 crc kubenswrapper[4843]: I0314 09:32:23.260790 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:32:23 crc kubenswrapper[4843]: I0314 09:32:23.349108 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2267c349-e72b-4476-bc83-0c80ccafb1fc" path="/var/lib/kubelet/pods/2267c349-e72b-4476-bc83-0c80ccafb1fc/volumes" Mar 14 09:32:24 crc kubenswrapper[4843]: I0314 09:32:24.054281 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerStarted","Data":"c0d8665e6c2da59a74f6081aa691475086398513074aff57bae2acbeed63b983"} Mar 14 09:32:24 crc kubenswrapper[4843]: I0314 09:32:24.054482 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerStarted","Data":"1d9cb778a5c6018784cd969ccf1e54bffc2d001ec2cdd2c538b4138790bcd1c4"} Mar 14 09:32:25 crc kubenswrapper[4843]: I0314 09:32:25.066256 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerStarted","Data":"84d2de8f9f78a9774751d0ae39ceb02c52c7688df02840a185f9247080f0b12f"} Mar 14 09:32:26 crc kubenswrapper[4843]: I0314 09:32:26.119449 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerStarted","Data":"d45adcb661ce46daaed762166c2c44da177aee39ee9ad8d74f77764714d1c8e2"} Mar 14 09:32:28 crc kubenswrapper[4843]: I0314 09:32:28.137925 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerStarted","Data":"4e77f996f65b0aeddd27d3c8d93056e06c4ba0e348778c82695b0e0ea8c48eca"} Mar 14 09:32:28 crc kubenswrapper[4843]: I0314 09:32:28.139786 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:32:28 crc kubenswrapper[4843]: I0314 09:32:28.174809 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.484732005 podStartE2EDuration="6.174791172s" podCreationTimestamp="2026-03-14 09:32:22 +0000 UTC" firstStartedPulling="2026-03-14 09:32:23.27466522 +0000 UTC m=+1310.587276338" lastFinishedPulling="2026-03-14 09:32:26.964724377 +0000 UTC m=+1314.277335505" observedRunningTime="2026-03-14 09:32:28.174251789 +0000 UTC m=+1315.486862927" watchObservedRunningTime="2026-03-14 09:32:28.174791172 +0000 UTC m=+1315.487402290" Mar 14 09:32:45 crc kubenswrapper[4843]: I0314 09:32:45.387836 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:32:45 crc kubenswrapper[4843]: I0314 09:32:45.388491 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:32:48 crc kubenswrapper[4843]: I0314 09:32:48.271401 4843 scope.go:117] "RemoveContainer" containerID="da3c83f872d4062855f44c6aa2a2ecdda7254b497fb9af510c75abca102cc7fe" Mar 14 09:32:48 crc kubenswrapper[4843]: I0314 09:32:48.296722 4843 scope.go:117] "RemoveContainer" containerID="f1e0749317b316f6dbc387640e4947fc71f55e4904687cd9e916c67ef42dfa1b" Mar 14 09:32:52 crc kubenswrapper[4843]: I0314 09:32:52.768438 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:33:15 crc kubenswrapper[4843]: I0314 09:33:15.387099 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:33:15 crc kubenswrapper[4843]: I0314 09:33:15.387722 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:33:15 crc kubenswrapper[4843]: I0314 09:33:15.387776 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:33:15 crc kubenswrapper[4843]: I0314 09:33:15.388517 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea69e0d05b7309b15d93d55c71a74bebdf19a7967db0d10155ee8a1a048179ea"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:33:15 crc kubenswrapper[4843]: I0314 09:33:15.388584 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://ea69e0d05b7309b15d93d55c71a74bebdf19a7967db0d10155ee8a1a048179ea" gracePeriod=600 Mar 14 09:33:16 crc kubenswrapper[4843]: I0314 09:33:16.517172 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="ea69e0d05b7309b15d93d55c71a74bebdf19a7967db0d10155ee8a1a048179ea" exitCode=0 Mar 14 09:33:16 crc kubenswrapper[4843]: I0314 09:33:16.517263 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"ea69e0d05b7309b15d93d55c71a74bebdf19a7967db0d10155ee8a1a048179ea"} Mar 14 09:33:16 crc kubenswrapper[4843]: I0314 09:33:16.517774 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0"} Mar 14 09:33:16 crc kubenswrapper[4843]: I0314 09:33:16.517807 4843 scope.go:117] "RemoveContainer" containerID="1c0b6bcfd5822f50fc94ed067fd20658758c736e480b5fd0ec55da8213db2e74" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.138714 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558014-svhvt"] Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.140405 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.143416 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.143739 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.143999 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.148584 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558014-svhvt"] Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.251315 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm58f\" (UniqueName: \"kubernetes.io/projected/8d2398db-31e4-49c5-bc8d-08157c7a3dbd-kube-api-access-fm58f\") pod \"auto-csr-approver-29558014-svhvt\" (UID: \"8d2398db-31e4-49c5-bc8d-08157c7a3dbd\") " pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.353657 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm58f\" (UniqueName: \"kubernetes.io/projected/8d2398db-31e4-49c5-bc8d-08157c7a3dbd-kube-api-access-fm58f\") pod \"auto-csr-approver-29558014-svhvt\" (UID: \"8d2398db-31e4-49c5-bc8d-08157c7a3dbd\") " pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.376730 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm58f\" (UniqueName: \"kubernetes.io/projected/8d2398db-31e4-49c5-bc8d-08157c7a3dbd-kube-api-access-fm58f\") pod \"auto-csr-approver-29558014-svhvt\" (UID: \"8d2398db-31e4-49c5-bc8d-08157c7a3dbd\") " pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.456467 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.873257 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558014-svhvt"] Mar 14 09:34:00 crc kubenswrapper[4843]: I0314 09:34:00.882146 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:34:01 crc kubenswrapper[4843]: I0314 09:34:01.906895 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558014-svhvt" event={"ID":"8d2398db-31e4-49c5-bc8d-08157c7a3dbd","Type":"ContainerStarted","Data":"c065e5c017c9dd7393dff16c7939463182274cc72b5804f02b665cb74d1259d6"} Mar 14 09:34:02 crc kubenswrapper[4843]: I0314 09:34:02.916360 4843 generic.go:334] "Generic (PLEG): container finished" podID="8d2398db-31e4-49c5-bc8d-08157c7a3dbd" containerID="984edb3276c5815657228407cea74e85d271d8b79c56c353c444ebabcc6a9664" exitCode=0 Mar 14 09:34:02 crc kubenswrapper[4843]: I0314 09:34:02.916409 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558014-svhvt" event={"ID":"8d2398db-31e4-49c5-bc8d-08157c7a3dbd","Type":"ContainerDied","Data":"984edb3276c5815657228407cea74e85d271d8b79c56c353c444ebabcc6a9664"} Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.219642 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.319930 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm58f\" (UniqueName: \"kubernetes.io/projected/8d2398db-31e4-49c5-bc8d-08157c7a3dbd-kube-api-access-fm58f\") pod \"8d2398db-31e4-49c5-bc8d-08157c7a3dbd\" (UID: \"8d2398db-31e4-49c5-bc8d-08157c7a3dbd\") " Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.324948 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2398db-31e4-49c5-bc8d-08157c7a3dbd-kube-api-access-fm58f" (OuterVolumeSpecName: "kube-api-access-fm58f") pod "8d2398db-31e4-49c5-bc8d-08157c7a3dbd" (UID: "8d2398db-31e4-49c5-bc8d-08157c7a3dbd"). InnerVolumeSpecName "kube-api-access-fm58f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.421523 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm58f\" (UniqueName: \"kubernetes.io/projected/8d2398db-31e4-49c5-bc8d-08157c7a3dbd-kube-api-access-fm58f\") on node \"crc\" DevicePath \"\"" Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.933658 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558014-svhvt" event={"ID":"8d2398db-31e4-49c5-bc8d-08157c7a3dbd","Type":"ContainerDied","Data":"c065e5c017c9dd7393dff16c7939463182274cc72b5804f02b665cb74d1259d6"} Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.933704 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c065e5c017c9dd7393dff16c7939463182274cc72b5804f02b665cb74d1259d6" Mar 14 09:34:04 crc kubenswrapper[4843]: I0314 09:34:04.933715 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558014-svhvt" Mar 14 09:34:05 crc kubenswrapper[4843]: I0314 09:34:05.305797 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558008-x58wg"] Mar 14 09:34:05 crc kubenswrapper[4843]: I0314 09:34:05.318904 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558008-x58wg"] Mar 14 09:34:05 crc kubenswrapper[4843]: I0314 09:34:05.362125 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e5343d1-8bed-448c-a87f-a032b4f6a1a2" path="/var/lib/kubelet/pods/6e5343d1-8bed-448c-a87f-a032b4f6a1a2/volumes" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.000528 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f559l"] Mar 14 09:34:07 crc kubenswrapper[4843]: E0314 09:34:07.001070 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2398db-31e4-49c5-bc8d-08157c7a3dbd" containerName="oc" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.001087 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2398db-31e4-49c5-bc8d-08157c7a3dbd" containerName="oc" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.001369 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2398db-31e4-49c5-bc8d-08157c7a3dbd" containerName="oc" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.003125 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.016149 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f559l"] Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.062010 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-utilities\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.062079 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-catalog-content\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.062113 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjdsm\" (UniqueName: \"kubernetes.io/projected/08198efb-0412-47ba-ad09-c2177752f649-kube-api-access-fjdsm\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.164971 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-utilities\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.165027 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-catalog-content\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.165058 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjdsm\" (UniqueName: \"kubernetes.io/projected/08198efb-0412-47ba-ad09-c2177752f649-kube-api-access-fjdsm\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.165516 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-catalog-content\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.165671 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-utilities\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.190015 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjdsm\" (UniqueName: \"kubernetes.io/projected/08198efb-0412-47ba-ad09-c2177752f649-kube-api-access-fjdsm\") pod \"redhat-operators-f559l\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.331352 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.793894 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f559l"] Mar 14 09:34:07 crc kubenswrapper[4843]: I0314 09:34:07.958153 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerStarted","Data":"05f014d45b0ca2791dfcee2ed2a70c0364728e10a6864586121120d051ced6d3"} Mar 14 09:34:08 crc kubenswrapper[4843]: I0314 09:34:08.968507 4843 generic.go:334] "Generic (PLEG): container finished" podID="08198efb-0412-47ba-ad09-c2177752f649" containerID="811ee9b975fffbd05e1e889c657a444ebb8e6676fe6c47045ce96f9bf23a28c4" exitCode=0 Mar 14 09:34:08 crc kubenswrapper[4843]: I0314 09:34:08.968812 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerDied","Data":"811ee9b975fffbd05e1e889c657a444ebb8e6676fe6c47045ce96f9bf23a28c4"} Mar 14 09:34:09 crc kubenswrapper[4843]: I0314 09:34:09.985813 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerStarted","Data":"9e4ed19f530dfc2214f8d7576fd1737af57831910f4c7a90e44bfc4e25ae6e67"} Mar 14 09:34:12 crc kubenswrapper[4843]: I0314 09:34:12.010872 4843 generic.go:334] "Generic (PLEG): container finished" podID="08198efb-0412-47ba-ad09-c2177752f649" containerID="9e4ed19f530dfc2214f8d7576fd1737af57831910f4c7a90e44bfc4e25ae6e67" exitCode=0 Mar 14 09:34:12 crc kubenswrapper[4843]: I0314 09:34:12.010973 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerDied","Data":"9e4ed19f530dfc2214f8d7576fd1737af57831910f4c7a90e44bfc4e25ae6e67"} Mar 14 09:34:13 crc kubenswrapper[4843]: I0314 09:34:13.022541 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerStarted","Data":"a3ac79e4077b43ec9e0e16749714f065ae9f1b982835ad8ce98eeb18ab720ef3"} Mar 14 09:34:13 crc kubenswrapper[4843]: I0314 09:34:13.041704 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f559l" podStartSLOduration=3.604003202 podStartE2EDuration="7.041681977s" podCreationTimestamp="2026-03-14 09:34:06 +0000 UTC" firstStartedPulling="2026-03-14 09:34:08.97101872 +0000 UTC m=+1416.283629868" lastFinishedPulling="2026-03-14 09:34:12.408697515 +0000 UTC m=+1419.721308643" observedRunningTime="2026-03-14 09:34:13.037056182 +0000 UTC m=+1420.349667300" watchObservedRunningTime="2026-03-14 09:34:13.041681977 +0000 UTC m=+1420.354293105" Mar 14 09:34:17 crc kubenswrapper[4843]: I0314 09:34:17.331597 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:17 crc kubenswrapper[4843]: I0314 09:34:17.331903 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:18 crc kubenswrapper[4843]: I0314 09:34:18.435765 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f559l" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="registry-server" probeResult="failure" output=< Mar 14 09:34:18 crc kubenswrapper[4843]: timeout: failed to connect service ":50051" within 1s Mar 14 09:34:18 crc kubenswrapper[4843]: > Mar 14 09:34:27 crc kubenswrapper[4843]: I0314 09:34:27.402947 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:27 crc kubenswrapper[4843]: I0314 09:34:27.462589 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:30 crc kubenswrapper[4843]: I0314 09:34:30.993977 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f559l"] Mar 14 09:34:30 crc kubenswrapper[4843]: I0314 09:34:30.994511 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f559l" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="registry-server" containerID="cri-o://a3ac79e4077b43ec9e0e16749714f065ae9f1b982835ad8ce98eeb18ab720ef3" gracePeriod=2 Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.170554 4843 generic.go:334] "Generic (PLEG): container finished" podID="08198efb-0412-47ba-ad09-c2177752f649" containerID="a3ac79e4077b43ec9e0e16749714f065ae9f1b982835ad8ce98eeb18ab720ef3" exitCode=0 Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.170602 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerDied","Data":"a3ac79e4077b43ec9e0e16749714f065ae9f1b982835ad8ce98eeb18ab720ef3"} Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.404485 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.459105 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjdsm\" (UniqueName: \"kubernetes.io/projected/08198efb-0412-47ba-ad09-c2177752f649-kube-api-access-fjdsm\") pod \"08198efb-0412-47ba-ad09-c2177752f649\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.459187 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-catalog-content\") pod \"08198efb-0412-47ba-ad09-c2177752f649\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.459386 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-utilities\") pod \"08198efb-0412-47ba-ad09-c2177752f649\" (UID: \"08198efb-0412-47ba-ad09-c2177752f649\") " Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.460395 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-utilities" (OuterVolumeSpecName: "utilities") pod "08198efb-0412-47ba-ad09-c2177752f649" (UID: "08198efb-0412-47ba-ad09-c2177752f649"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.464963 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08198efb-0412-47ba-ad09-c2177752f649-kube-api-access-fjdsm" (OuterVolumeSpecName: "kube-api-access-fjdsm") pod "08198efb-0412-47ba-ad09-c2177752f649" (UID: "08198efb-0412-47ba-ad09-c2177752f649"). InnerVolumeSpecName "kube-api-access-fjdsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.560786 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.560821 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjdsm\" (UniqueName: \"kubernetes.io/projected/08198efb-0412-47ba-ad09-c2177752f649-kube-api-access-fjdsm\") on node \"crc\" DevicePath \"\"" Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.579864 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08198efb-0412-47ba-ad09-c2177752f649" (UID: "08198efb-0412-47ba-ad09-c2177752f649"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:34:31 crc kubenswrapper[4843]: I0314 09:34:31.662155 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08198efb-0412-47ba-ad09-c2177752f649-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.182226 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f559l" event={"ID":"08198efb-0412-47ba-ad09-c2177752f649","Type":"ContainerDied","Data":"05f014d45b0ca2791dfcee2ed2a70c0364728e10a6864586121120d051ced6d3"} Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.182328 4843 scope.go:117] "RemoveContainer" containerID="a3ac79e4077b43ec9e0e16749714f065ae9f1b982835ad8ce98eeb18ab720ef3" Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.182339 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f559l" Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.221047 4843 scope.go:117] "RemoveContainer" containerID="9e4ed19f530dfc2214f8d7576fd1737af57831910f4c7a90e44bfc4e25ae6e67" Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.221390 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f559l"] Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.230912 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f559l"] Mar 14 09:34:32 crc kubenswrapper[4843]: I0314 09:34:32.254768 4843 scope.go:117] "RemoveContainer" containerID="811ee9b975fffbd05e1e889c657a444ebb8e6676fe6c47045ce96f9bf23a28c4" Mar 14 09:34:33 crc kubenswrapper[4843]: I0314 09:34:33.360385 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08198efb-0412-47ba-ad09-c2177752f649" path="/var/lib/kubelet/pods/08198efb-0412-47ba-ad09-c2177752f649/volumes" Mar 14 09:34:48 crc kubenswrapper[4843]: I0314 09:34:48.409000 4843 scope.go:117] "RemoveContainer" containerID="84f73c8842bd583d796d0e6de35d870485969f1969e38951dc0297f05c21d115" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.404122 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5fntq"] Mar 14 09:34:51 crc kubenswrapper[4843]: E0314 09:34:51.405416 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="extract-content" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.405440 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="extract-content" Mar 14 09:34:51 crc kubenswrapper[4843]: E0314 09:34:51.405472 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="extract-utilities" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.405485 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="extract-utilities" Mar 14 09:34:51 crc kubenswrapper[4843]: E0314 09:34:51.405533 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="registry-server" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.405545 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="registry-server" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.405826 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="08198efb-0412-47ba-ad09-c2177752f649" containerName="registry-server" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.407607 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.412657 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5fntq"] Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.578222 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5m6w\" (UniqueName: \"kubernetes.io/projected/09a3d870-dac0-4b32-8889-23e2a17984f1-kube-api-access-g5m6w\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.578575 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-catalog-content\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.578706 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-utilities\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.680597 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5m6w\" (UniqueName: \"kubernetes.io/projected/09a3d870-dac0-4b32-8889-23e2a17984f1-kube-api-access-g5m6w\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.680674 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-catalog-content\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.680704 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-utilities\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.681211 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-catalog-content\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.681249 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-utilities\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.703655 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5m6w\" (UniqueName: \"kubernetes.io/projected/09a3d870-dac0-4b32-8889-23e2a17984f1-kube-api-access-g5m6w\") pod \"community-operators-5fntq\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:51 crc kubenswrapper[4843]: I0314 09:34:51.729351 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:34:52 crc kubenswrapper[4843]: I0314 09:34:52.234910 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5fntq"] Mar 14 09:34:52 crc kubenswrapper[4843]: I0314 09:34:52.352833 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerStarted","Data":"ad80bd1279ffbcfd3e89f496f6813cd0361b1eea4a7aed77b0fbd01d2e1a1250"} Mar 14 09:34:53 crc kubenswrapper[4843]: I0314 09:34:53.365162 4843 generic.go:334] "Generic (PLEG): container finished" podID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerID="c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91" exitCode=0 Mar 14 09:34:53 crc kubenswrapper[4843]: I0314 09:34:53.365233 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerDied","Data":"c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91"} Mar 14 09:34:54 crc kubenswrapper[4843]: I0314 09:34:54.379829 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerStarted","Data":"a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2"} Mar 14 09:34:55 crc kubenswrapper[4843]: I0314 09:34:55.388324 4843 generic.go:334] "Generic (PLEG): container finished" podID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerID="a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2" exitCode=0 Mar 14 09:34:55 crc kubenswrapper[4843]: I0314 09:34:55.388372 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerDied","Data":"a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2"} Mar 14 09:34:56 crc kubenswrapper[4843]: I0314 09:34:56.397224 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerStarted","Data":"75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e"} Mar 14 09:34:56 crc kubenswrapper[4843]: I0314 09:34:56.413521 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5fntq" podStartSLOduration=2.930924223 podStartE2EDuration="5.413500131s" podCreationTimestamp="2026-03-14 09:34:51 +0000 UTC" firstStartedPulling="2026-03-14 09:34:53.369636055 +0000 UTC m=+1460.682247183" lastFinishedPulling="2026-03-14 09:34:55.852211963 +0000 UTC m=+1463.164823091" observedRunningTime="2026-03-14 09:34:56.41184497 +0000 UTC m=+1463.724456098" watchObservedRunningTime="2026-03-14 09:34:56.413500131 +0000 UTC m=+1463.726111259" Mar 14 09:35:01 crc kubenswrapper[4843]: I0314 09:35:01.729707 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:35:01 crc kubenswrapper[4843]: I0314 09:35:01.731441 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:35:01 crc kubenswrapper[4843]: I0314 09:35:01.778973 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:35:02 crc kubenswrapper[4843]: I0314 09:35:02.556721 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.387100 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5fntq"] Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.507925 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5fntq" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="registry-server" containerID="cri-o://75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e" gracePeriod=2 Mar 14 09:35:05 crc kubenswrapper[4843]: E0314 09:35:05.740986 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a3d870_dac0_4b32_8889_23e2a17984f1.slice/crio-conmon-75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e.scope\": RecentStats: unable to find data in memory cache]" Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.922326 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.932913 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5m6w\" (UniqueName: \"kubernetes.io/projected/09a3d870-dac0-4b32-8889-23e2a17984f1-kube-api-access-g5m6w\") pod \"09a3d870-dac0-4b32-8889-23e2a17984f1\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.932977 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-utilities\") pod \"09a3d870-dac0-4b32-8889-23e2a17984f1\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.933055 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-catalog-content\") pod \"09a3d870-dac0-4b32-8889-23e2a17984f1\" (UID: \"09a3d870-dac0-4b32-8889-23e2a17984f1\") " Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.934037 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-utilities" (OuterVolumeSpecName: "utilities") pod "09a3d870-dac0-4b32-8889-23e2a17984f1" (UID: "09a3d870-dac0-4b32-8889-23e2a17984f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:35:05 crc kubenswrapper[4843]: I0314 09:35:05.938566 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a3d870-dac0-4b32-8889-23e2a17984f1-kube-api-access-g5m6w" (OuterVolumeSpecName: "kube-api-access-g5m6w") pod "09a3d870-dac0-4b32-8889-23e2a17984f1" (UID: "09a3d870-dac0-4b32-8889-23e2a17984f1"). InnerVolumeSpecName "kube-api-access-g5m6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.035035 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5m6w\" (UniqueName: \"kubernetes.io/projected/09a3d870-dac0-4b32-8889-23e2a17984f1-kube-api-access-g5m6w\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.035072 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.077695 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09a3d870-dac0-4b32-8889-23e2a17984f1" (UID: "09a3d870-dac0-4b32-8889-23e2a17984f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.136851 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a3d870-dac0-4b32-8889-23e2a17984f1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.517881 4843 generic.go:334] "Generic (PLEG): container finished" podID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerID="75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e" exitCode=0 Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.517917 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerDied","Data":"75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e"} Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.517984 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5fntq" event={"ID":"09a3d870-dac0-4b32-8889-23e2a17984f1","Type":"ContainerDied","Data":"ad80bd1279ffbcfd3e89f496f6813cd0361b1eea4a7aed77b0fbd01d2e1a1250"} Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.518004 4843 scope.go:117] "RemoveContainer" containerID="75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.517997 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5fntq" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.545382 4843 scope.go:117] "RemoveContainer" containerID="a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.559758 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5fntq"] Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.568020 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5fntq"] Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.576824 4843 scope.go:117] "RemoveContainer" containerID="c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.603240 4843 scope.go:117] "RemoveContainer" containerID="75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e" Mar 14 09:35:06 crc kubenswrapper[4843]: E0314 09:35:06.603675 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e\": container with ID starting with 75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e not found: ID does not exist" containerID="75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.603733 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e"} err="failed to get container status \"75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e\": rpc error: code = NotFound desc = could not find container \"75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e\": container with ID starting with 75e86a3207525e258aa19934b332977f7a1865007b4f347a68db5ec414605d4e not found: ID does not exist" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.603757 4843 scope.go:117] "RemoveContainer" containerID="a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2" Mar 14 09:35:06 crc kubenswrapper[4843]: E0314 09:35:06.603978 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2\": container with ID starting with a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2 not found: ID does not exist" containerID="a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.604006 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2"} err="failed to get container status \"a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2\": rpc error: code = NotFound desc = could not find container \"a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2\": container with ID starting with a7ab35c42232cb4536e0738b047983363e075c8738369de03d8636101dd254f2 not found: ID does not exist" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.604023 4843 scope.go:117] "RemoveContainer" containerID="c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91" Mar 14 09:35:06 crc kubenswrapper[4843]: E0314 09:35:06.605257 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91\": container with ID starting with c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91 not found: ID does not exist" containerID="c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91" Mar 14 09:35:06 crc kubenswrapper[4843]: I0314 09:35:06.605318 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91"} err="failed to get container status \"c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91\": rpc error: code = NotFound desc = could not find container \"c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91\": container with ID starting with c279e01a6613185edb65ddc5d3939d2c8659c7badc8a84b0f6af88ced3f09e91 not found: ID does not exist" Mar 14 09:35:07 crc kubenswrapper[4843]: I0314 09:35:07.348011 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" path="/var/lib/kubelet/pods/09a3d870-dac0-4b32-8889-23e2a17984f1/volumes" Mar 14 09:35:15 crc kubenswrapper[4843]: I0314 09:35:15.387034 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:35:15 crc kubenswrapper[4843]: I0314 09:35:15.387657 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.598026 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d42w5"] Mar 14 09:35:19 crc kubenswrapper[4843]: E0314 09:35:19.598953 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="registry-server" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.598969 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="registry-server" Mar 14 09:35:19 crc kubenswrapper[4843]: E0314 09:35:19.598985 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="extract-utilities" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.598992 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="extract-utilities" Mar 14 09:35:19 crc kubenswrapper[4843]: E0314 09:35:19.599022 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="extract-content" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.599029 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="extract-content" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.599221 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a3d870-dac0-4b32-8889-23e2a17984f1" containerName="registry-server" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.600580 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.605616 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d42w5"] Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.657821 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-utilities\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.657886 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthzf\" (UniqueName: \"kubernetes.io/projected/74391f36-83e2-4ca8-b348-28eff316f9ae-kube-api-access-kthzf\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.657914 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-catalog-content\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.759028 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-utilities\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.759116 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthzf\" (UniqueName: \"kubernetes.io/projected/74391f36-83e2-4ca8-b348-28eff316f9ae-kube-api-access-kthzf\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.759153 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-catalog-content\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.759554 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-utilities\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.759637 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-catalog-content\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.778190 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthzf\" (UniqueName: \"kubernetes.io/projected/74391f36-83e2-4ca8-b348-28eff316f9ae-kube-api-access-kthzf\") pod \"certified-operators-d42w5\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:19 crc kubenswrapper[4843]: I0314 09:35:19.921167 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:20 crc kubenswrapper[4843]: I0314 09:35:20.434651 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d42w5"] Mar 14 09:35:20 crc kubenswrapper[4843]: I0314 09:35:20.624620 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerStarted","Data":"b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c"} Mar 14 09:35:20 crc kubenswrapper[4843]: I0314 09:35:20.625078 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerStarted","Data":"ff1660f0eb34e4e95bdfc62fd5518f2c9b8e229fa7b2d3e058fed573f6dc3e2e"} Mar 14 09:35:21 crc kubenswrapper[4843]: I0314 09:35:21.635340 4843 generic.go:334] "Generic (PLEG): container finished" podID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerID="b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c" exitCode=0 Mar 14 09:35:21 crc kubenswrapper[4843]: I0314 09:35:21.635379 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerDied","Data":"b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c"} Mar 14 09:35:22 crc kubenswrapper[4843]: I0314 09:35:22.645221 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerStarted","Data":"092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9"} Mar 14 09:35:23 crc kubenswrapper[4843]: I0314 09:35:23.657459 4843 generic.go:334] "Generic (PLEG): container finished" podID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerID="092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9" exitCode=0 Mar 14 09:35:23 crc kubenswrapper[4843]: I0314 09:35:23.657571 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerDied","Data":"092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9"} Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.667437 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerStarted","Data":"0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d"} Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.688961 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d42w5" podStartSLOduration=3.242432195 podStartE2EDuration="5.688939366s" podCreationTimestamp="2026-03-14 09:35:19 +0000 UTC" firstStartedPulling="2026-03-14 09:35:21.639364798 +0000 UTC m=+1488.951975926" lastFinishedPulling="2026-03-14 09:35:24.085871969 +0000 UTC m=+1491.398483097" observedRunningTime="2026-03-14 09:35:24.683746707 +0000 UTC m=+1491.996357825" watchObservedRunningTime="2026-03-14 09:35:24.688939366 +0000 UTC m=+1492.001550494" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.800894 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dhjfb"] Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.803025 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.811811 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dhjfb"] Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.861806 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-catalog-content\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.861922 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-utilities\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.862013 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6cqf\" (UniqueName: \"kubernetes.io/projected/77464f0c-656a-456c-894e-87f708878520-kube-api-access-f6cqf\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.963231 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6cqf\" (UniqueName: \"kubernetes.io/projected/77464f0c-656a-456c-894e-87f708878520-kube-api-access-f6cqf\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.963339 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-catalog-content\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.963395 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-utilities\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.963838 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-catalog-content\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.963882 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-utilities\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:24 crc kubenswrapper[4843]: I0314 09:35:24.984005 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6cqf\" (UniqueName: \"kubernetes.io/projected/77464f0c-656a-456c-894e-87f708878520-kube-api-access-f6cqf\") pod \"redhat-marketplace-dhjfb\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:25 crc kubenswrapper[4843]: I0314 09:35:25.125242 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:25 crc kubenswrapper[4843]: I0314 09:35:25.619211 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dhjfb"] Mar 14 09:35:25 crc kubenswrapper[4843]: W0314 09:35:25.629922 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77464f0c_656a_456c_894e_87f708878520.slice/crio-c62a2c8356024fb4def68ff1fcc14001133e33a51962a721cdc9a7df9c4f7fdf WatchSource:0}: Error finding container c62a2c8356024fb4def68ff1fcc14001133e33a51962a721cdc9a7df9c4f7fdf: Status 404 returned error can't find the container with id c62a2c8356024fb4def68ff1fcc14001133e33a51962a721cdc9a7df9c4f7fdf Mar 14 09:35:25 crc kubenswrapper[4843]: I0314 09:35:25.677364 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dhjfb" event={"ID":"77464f0c-656a-456c-894e-87f708878520","Type":"ContainerStarted","Data":"c62a2c8356024fb4def68ff1fcc14001133e33a51962a721cdc9a7df9c4f7fdf"} Mar 14 09:35:26 crc kubenswrapper[4843]: I0314 09:35:26.696002 4843 generic.go:334] "Generic (PLEG): container finished" podID="77464f0c-656a-456c-894e-87f708878520" containerID="1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f" exitCode=0 Mar 14 09:35:26 crc kubenswrapper[4843]: I0314 09:35:26.696334 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dhjfb" event={"ID":"77464f0c-656a-456c-894e-87f708878520","Type":"ContainerDied","Data":"1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f"} Mar 14 09:35:27 crc kubenswrapper[4843]: I0314 09:35:27.706945 4843 generic.go:334] "Generic (PLEG): container finished" podID="77464f0c-656a-456c-894e-87f708878520" containerID="88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84" exitCode=0 Mar 14 09:35:27 crc kubenswrapper[4843]: I0314 09:35:27.707011 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dhjfb" event={"ID":"77464f0c-656a-456c-894e-87f708878520","Type":"ContainerDied","Data":"88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84"} Mar 14 09:35:28 crc kubenswrapper[4843]: I0314 09:35:28.717719 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dhjfb" event={"ID":"77464f0c-656a-456c-894e-87f708878520","Type":"ContainerStarted","Data":"36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62"} Mar 14 09:35:28 crc kubenswrapper[4843]: I0314 09:35:28.748886 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dhjfb" podStartSLOduration=3.27945982 podStartE2EDuration="4.748865805s" podCreationTimestamp="2026-03-14 09:35:24 +0000 UTC" firstStartedPulling="2026-03-14 09:35:26.700163274 +0000 UTC m=+1494.012774412" lastFinishedPulling="2026-03-14 09:35:28.169569259 +0000 UTC m=+1495.482180397" observedRunningTime="2026-03-14 09:35:28.741333198 +0000 UTC m=+1496.053944346" watchObservedRunningTime="2026-03-14 09:35:28.748865805 +0000 UTC m=+1496.061476933" Mar 14 09:35:29 crc kubenswrapper[4843]: I0314 09:35:29.921916 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:29 crc kubenswrapper[4843]: I0314 09:35:29.922052 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:29 crc kubenswrapper[4843]: I0314 09:35:29.969433 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:30 crc kubenswrapper[4843]: I0314 09:35:30.788262 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:34 crc kubenswrapper[4843]: I0314 09:35:34.788912 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d42w5"] Mar 14 09:35:34 crc kubenswrapper[4843]: I0314 09:35:34.789497 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d42w5" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="registry-server" containerID="cri-o://0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d" gracePeriod=2 Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.125872 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.126212 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.171569 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.347842 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.382582 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-catalog-content\") pod \"74391f36-83e2-4ca8-b348-28eff316f9ae\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.382957 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kthzf\" (UniqueName: \"kubernetes.io/projected/74391f36-83e2-4ca8-b348-28eff316f9ae-kube-api-access-kthzf\") pod \"74391f36-83e2-4ca8-b348-28eff316f9ae\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.383050 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-utilities\") pod \"74391f36-83e2-4ca8-b348-28eff316f9ae\" (UID: \"74391f36-83e2-4ca8-b348-28eff316f9ae\") " Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.383839 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-utilities" (OuterVolumeSpecName: "utilities") pod "74391f36-83e2-4ca8-b348-28eff316f9ae" (UID: "74391f36-83e2-4ca8-b348-28eff316f9ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.388614 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74391f36-83e2-4ca8-b348-28eff316f9ae-kube-api-access-kthzf" (OuterVolumeSpecName: "kube-api-access-kthzf") pod "74391f36-83e2-4ca8-b348-28eff316f9ae" (UID: "74391f36-83e2-4ca8-b348-28eff316f9ae"). InnerVolumeSpecName "kube-api-access-kthzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.443443 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74391f36-83e2-4ca8-b348-28eff316f9ae" (UID: "74391f36-83e2-4ca8-b348-28eff316f9ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.484554 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.484590 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kthzf\" (UniqueName: \"kubernetes.io/projected/74391f36-83e2-4ca8-b348-28eff316f9ae-kube-api-access-kthzf\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.484605 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391f36-83e2-4ca8-b348-28eff316f9ae-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.771456 4843 generic.go:334] "Generic (PLEG): container finished" podID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerID="0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d" exitCode=0 Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.771522 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerDied","Data":"0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d"} Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.771792 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d42w5" event={"ID":"74391f36-83e2-4ca8-b348-28eff316f9ae","Type":"ContainerDied","Data":"ff1660f0eb34e4e95bdfc62fd5518f2c9b8e229fa7b2d3e058fed573f6dc3e2e"} Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.771828 4843 scope.go:117] "RemoveContainer" containerID="0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.771536 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d42w5" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.809088 4843 scope.go:117] "RemoveContainer" containerID="092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.823787 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d42w5"] Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.827149 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d42w5"] Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.827846 4843 scope.go:117] "RemoveContainer" containerID="b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.829814 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.868619 4843 scope.go:117] "RemoveContainer" containerID="0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d" Mar 14 09:35:35 crc kubenswrapper[4843]: E0314 09:35:35.869066 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d\": container with ID starting with 0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d not found: ID does not exist" containerID="0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.869115 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d"} err="failed to get container status \"0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d\": rpc error: code = NotFound desc = could not find container \"0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d\": container with ID starting with 0dd7af6be81eac5b34f093c43b84addd15f7c0df39e58d78e91c8fda3e19148d not found: ID does not exist" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.869141 4843 scope.go:117] "RemoveContainer" containerID="092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9" Mar 14 09:35:35 crc kubenswrapper[4843]: E0314 09:35:35.869521 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9\": container with ID starting with 092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9 not found: ID does not exist" containerID="092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.869577 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9"} err="failed to get container status \"092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9\": rpc error: code = NotFound desc = could not find container \"092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9\": container with ID starting with 092e4dead1d93d655a94a147910c5985bbd8531d2556832d448805f03254d4e9 not found: ID does not exist" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.869591 4843 scope.go:117] "RemoveContainer" containerID="b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c" Mar 14 09:35:35 crc kubenswrapper[4843]: E0314 09:35:35.869877 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c\": container with ID starting with b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c not found: ID does not exist" containerID="b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c" Mar 14 09:35:35 crc kubenswrapper[4843]: I0314 09:35:35.869909 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c"} err="failed to get container status \"b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c\": rpc error: code = NotFound desc = could not find container \"b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c\": container with ID starting with b68378d26e5c630ce3d91fdb822078fccce9ab7a906088d6b6a05eb8a4f2c30c not found: ID does not exist" Mar 14 09:35:37 crc kubenswrapper[4843]: I0314 09:35:37.349226 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" path="/var/lib/kubelet/pods/74391f36-83e2-4ca8-b348-28eff316f9ae/volumes" Mar 14 09:35:39 crc kubenswrapper[4843]: I0314 09:35:39.988213 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dhjfb"] Mar 14 09:35:39 crc kubenswrapper[4843]: I0314 09:35:39.989343 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dhjfb" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="registry-server" containerID="cri-o://36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62" gracePeriod=2 Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.492964 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.564462 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6cqf\" (UniqueName: \"kubernetes.io/projected/77464f0c-656a-456c-894e-87f708878520-kube-api-access-f6cqf\") pod \"77464f0c-656a-456c-894e-87f708878520\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.564514 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-catalog-content\") pod \"77464f0c-656a-456c-894e-87f708878520\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.564553 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-utilities\") pod \"77464f0c-656a-456c-894e-87f708878520\" (UID: \"77464f0c-656a-456c-894e-87f708878520\") " Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.565529 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-utilities" (OuterVolumeSpecName: "utilities") pod "77464f0c-656a-456c-894e-87f708878520" (UID: "77464f0c-656a-456c-894e-87f708878520"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.569063 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77464f0c-656a-456c-894e-87f708878520-kube-api-access-f6cqf" (OuterVolumeSpecName: "kube-api-access-f6cqf") pod "77464f0c-656a-456c-894e-87f708878520" (UID: "77464f0c-656a-456c-894e-87f708878520"). InnerVolumeSpecName "kube-api-access-f6cqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.590970 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77464f0c-656a-456c-894e-87f708878520" (UID: "77464f0c-656a-456c-894e-87f708878520"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.665544 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6cqf\" (UniqueName: \"kubernetes.io/projected/77464f0c-656a-456c-894e-87f708878520-kube-api-access-f6cqf\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.665578 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.665588 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77464f0c-656a-456c-894e-87f708878520-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.841574 4843 generic.go:334] "Generic (PLEG): container finished" podID="77464f0c-656a-456c-894e-87f708878520" containerID="36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62" exitCode=0 Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.841612 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dhjfb" event={"ID":"77464f0c-656a-456c-894e-87f708878520","Type":"ContainerDied","Data":"36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62"} Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.841657 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dhjfb" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.841668 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dhjfb" event={"ID":"77464f0c-656a-456c-894e-87f708878520","Type":"ContainerDied","Data":"c62a2c8356024fb4def68ff1fcc14001133e33a51962a721cdc9a7df9c4f7fdf"} Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.841691 4843 scope.go:117] "RemoveContainer" containerID="36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.868878 4843 scope.go:117] "RemoveContainer" containerID="88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.878073 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dhjfb"] Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.884052 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dhjfb"] Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.905605 4843 scope.go:117] "RemoveContainer" containerID="1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.925668 4843 scope.go:117] "RemoveContainer" containerID="36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62" Mar 14 09:35:40 crc kubenswrapper[4843]: E0314 09:35:40.926213 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62\": container with ID starting with 36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62 not found: ID does not exist" containerID="36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.926255 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62"} err="failed to get container status \"36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62\": rpc error: code = NotFound desc = could not find container \"36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62\": container with ID starting with 36f75fe02d534075e1bb7f1cccfb747eab1f771a2a22743aaea9de8f06349c62 not found: ID does not exist" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.926305 4843 scope.go:117] "RemoveContainer" containerID="88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84" Mar 14 09:35:40 crc kubenswrapper[4843]: E0314 09:35:40.926755 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84\": container with ID starting with 88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84 not found: ID does not exist" containerID="88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.926781 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84"} err="failed to get container status \"88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84\": rpc error: code = NotFound desc = could not find container \"88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84\": container with ID starting with 88073a35972222b160ec2fb1bd698cab524213bed8fc74661f96de3d0837dc84 not found: ID does not exist" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.926795 4843 scope.go:117] "RemoveContainer" containerID="1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f" Mar 14 09:35:40 crc kubenswrapper[4843]: E0314 09:35:40.927203 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f\": container with ID starting with 1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f not found: ID does not exist" containerID="1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f" Mar 14 09:35:40 crc kubenswrapper[4843]: I0314 09:35:40.927235 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f"} err="failed to get container status \"1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f\": rpc error: code = NotFound desc = could not find container \"1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f\": container with ID starting with 1ee749cd817e787d5d6b6d207f764c20c6ee3f01da119115da494871e55a397f not found: ID does not exist" Mar 14 09:35:41 crc kubenswrapper[4843]: I0314 09:35:41.352842 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77464f0c-656a-456c-894e-87f708878520" path="/var/lib/kubelet/pods/77464f0c-656a-456c-894e-87f708878520/volumes" Mar 14 09:35:45 crc kubenswrapper[4843]: I0314 09:35:45.388078 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:35:45 crc kubenswrapper[4843]: I0314 09:35:45.388486 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.498620 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558016-9r6w4"] Mar 14 09:36:01 crc kubenswrapper[4843]: E0314 09:36:01.499513 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="extract-content" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499527 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="extract-content" Mar 14 09:36:01 crc kubenswrapper[4843]: E0314 09:36:01.499539 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="registry-server" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499546 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="registry-server" Mar 14 09:36:01 crc kubenswrapper[4843]: E0314 09:36:01.499555 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="extract-content" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499561 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="extract-content" Mar 14 09:36:01 crc kubenswrapper[4843]: E0314 09:36:01.499577 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="registry-server" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499582 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="registry-server" Mar 14 09:36:01 crc kubenswrapper[4843]: E0314 09:36:01.499604 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="extract-utilities" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499609 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="extract-utilities" Mar 14 09:36:01 crc kubenswrapper[4843]: E0314 09:36:01.499623 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="extract-utilities" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499629 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="extract-utilities" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499768 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="74391f36-83e2-4ca8-b348-28eff316f9ae" containerName="registry-server" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.499784 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="77464f0c-656a-456c-894e-87f708878520" containerName="registry-server" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.500514 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.504929 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.505012 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.505816 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.527875 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558016-9r6w4"] Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.558554 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcz9r\" (UniqueName: \"kubernetes.io/projected/43022031-bb5c-498e-8ddb-933b3058a45f-kube-api-access-hcz9r\") pod \"auto-csr-approver-29558016-9r6w4\" (UID: \"43022031-bb5c-498e-8ddb-933b3058a45f\") " pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.660738 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcz9r\" (UniqueName: \"kubernetes.io/projected/43022031-bb5c-498e-8ddb-933b3058a45f-kube-api-access-hcz9r\") pod \"auto-csr-approver-29558016-9r6w4\" (UID: \"43022031-bb5c-498e-8ddb-933b3058a45f\") " pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.682468 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcz9r\" (UniqueName: \"kubernetes.io/projected/43022031-bb5c-498e-8ddb-933b3058a45f-kube-api-access-hcz9r\") pod \"auto-csr-approver-29558016-9r6w4\" (UID: \"43022031-bb5c-498e-8ddb-933b3058a45f\") " pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:01 crc kubenswrapper[4843]: I0314 09:36:01.826491 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:02 crc kubenswrapper[4843]: I0314 09:36:02.277862 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558016-9r6w4"] Mar 14 09:36:02 crc kubenswrapper[4843]: I0314 09:36:02.494174 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" event={"ID":"43022031-bb5c-498e-8ddb-933b3058a45f","Type":"ContainerStarted","Data":"14290503645f4e7c146b4c2577b0144a504325d8d4f378d6eaeb20fccba1dba0"} Mar 14 09:36:03 crc kubenswrapper[4843]: I0314 09:36:03.502107 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" event={"ID":"43022031-bb5c-498e-8ddb-933b3058a45f","Type":"ContainerStarted","Data":"db9356172cd17aa975a4dd73199e3644929a4395cd0dcfd07644f61db70898db"} Mar 14 09:36:03 crc kubenswrapper[4843]: I0314 09:36:03.523241 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" podStartSLOduration=2.732083485 podStartE2EDuration="3.523219652s" podCreationTimestamp="2026-03-14 09:36:00 +0000 UTC" firstStartedPulling="2026-03-14 09:36:02.284079666 +0000 UTC m=+1529.596690794" lastFinishedPulling="2026-03-14 09:36:03.075215833 +0000 UTC m=+1530.387826961" observedRunningTime="2026-03-14 09:36:03.513581283 +0000 UTC m=+1530.826192411" watchObservedRunningTime="2026-03-14 09:36:03.523219652 +0000 UTC m=+1530.835830780" Mar 14 09:36:04 crc kubenswrapper[4843]: I0314 09:36:04.509974 4843 generic.go:334] "Generic (PLEG): container finished" podID="43022031-bb5c-498e-8ddb-933b3058a45f" containerID="db9356172cd17aa975a4dd73199e3644929a4395cd0dcfd07644f61db70898db" exitCode=0 Mar 14 09:36:04 crc kubenswrapper[4843]: I0314 09:36:04.510040 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" event={"ID":"43022031-bb5c-498e-8ddb-933b3058a45f","Type":"ContainerDied","Data":"db9356172cd17aa975a4dd73199e3644929a4395cd0dcfd07644f61db70898db"} Mar 14 09:36:05 crc kubenswrapper[4843]: I0314 09:36:05.894652 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.021829 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcz9r\" (UniqueName: \"kubernetes.io/projected/43022031-bb5c-498e-8ddb-933b3058a45f-kube-api-access-hcz9r\") pod \"43022031-bb5c-498e-8ddb-933b3058a45f\" (UID: \"43022031-bb5c-498e-8ddb-933b3058a45f\") " Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.032002 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43022031-bb5c-498e-8ddb-933b3058a45f-kube-api-access-hcz9r" (OuterVolumeSpecName: "kube-api-access-hcz9r") pod "43022031-bb5c-498e-8ddb-933b3058a45f" (UID: "43022031-bb5c-498e-8ddb-933b3058a45f"). InnerVolumeSpecName "kube-api-access-hcz9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.123789 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcz9r\" (UniqueName: \"kubernetes.io/projected/43022031-bb5c-498e-8ddb-933b3058a45f-kube-api-access-hcz9r\") on node \"crc\" DevicePath \"\"" Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.434227 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558010-822zk"] Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.443542 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558010-822zk"] Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.524481 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" event={"ID":"43022031-bb5c-498e-8ddb-933b3058a45f","Type":"ContainerDied","Data":"14290503645f4e7c146b4c2577b0144a504325d8d4f378d6eaeb20fccba1dba0"} Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.524769 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14290503645f4e7c146b4c2577b0144a504325d8d4f378d6eaeb20fccba1dba0" Mar 14 09:36:06 crc kubenswrapper[4843]: I0314 09:36:06.524662 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558016-9r6w4" Mar 14 09:36:07 crc kubenswrapper[4843]: I0314 09:36:07.347902 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b" path="/var/lib/kubelet/pods/c7cc93ba-b787-4d2e-9bc1-6f9d003ccd9b/volumes" Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.387886 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.389169 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.389254 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.390355 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.390421 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" gracePeriod=600 Mar 14 09:36:15 crc kubenswrapper[4843]: E0314 09:36:15.511569 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.602876 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" exitCode=0 Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.602932 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0"} Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.602965 4843 scope.go:117] "RemoveContainer" containerID="ea69e0d05b7309b15d93d55c71a74bebdf19a7967db0d10155ee8a1a048179ea" Mar 14 09:36:15 crc kubenswrapper[4843]: I0314 09:36:15.603596 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:36:15 crc kubenswrapper[4843]: E0314 09:36:15.603876 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:36:27 crc kubenswrapper[4843]: I0314 09:36:27.338981 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:36:27 crc kubenswrapper[4843]: E0314 09:36:27.339782 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:36:41 crc kubenswrapper[4843]: I0314 09:36:41.339089 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:36:41 crc kubenswrapper[4843]: E0314 09:36:41.339942 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:36:48 crc kubenswrapper[4843]: I0314 09:36:48.580033 4843 scope.go:117] "RemoveContainer" containerID="60483461e6971b935c938888e01f74fbff2ad70a71ad49257931eae9c5033cc7" Mar 14 09:36:48 crc kubenswrapper[4843]: I0314 09:36:48.637763 4843 scope.go:117] "RemoveContainer" containerID="73576ee29aa9d69d81deb018b6ceaab25d7626e7f004309e239637860159d359" Mar 14 09:36:56 crc kubenswrapper[4843]: I0314 09:36:56.339419 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:36:56 crc kubenswrapper[4843]: E0314 09:36:56.340469 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:37:08 crc kubenswrapper[4843]: I0314 09:37:08.339371 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:37:08 crc kubenswrapper[4843]: E0314 09:37:08.341541 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:37:20 crc kubenswrapper[4843]: I0314 09:37:20.338464 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:37:20 crc kubenswrapper[4843]: E0314 09:37:20.340289 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:37:31 crc kubenswrapper[4843]: I0314 09:37:31.339397 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:37:31 crc kubenswrapper[4843]: E0314 09:37:31.340566 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:37:44 crc kubenswrapper[4843]: I0314 09:37:44.340068 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:37:44 crc kubenswrapper[4843]: E0314 09:37:44.341641 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:37:57 crc kubenswrapper[4843]: I0314 09:37:57.342005 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:37:57 crc kubenswrapper[4843]: E0314 09:37:57.342778 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.157193 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558018-kk95p"] Mar 14 09:38:00 crc kubenswrapper[4843]: E0314 09:38:00.158981 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43022031-bb5c-498e-8ddb-933b3058a45f" containerName="oc" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.159244 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="43022031-bb5c-498e-8ddb-933b3058a45f" containerName="oc" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.159490 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="43022031-bb5c-498e-8ddb-933b3058a45f" containerName="oc" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.160112 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.162492 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.162626 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.164366 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.170172 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558018-kk95p"] Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.336676 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bj2r\" (UniqueName: \"kubernetes.io/projected/5c586de6-9329-471d-b657-11668068ed0f-kube-api-access-8bj2r\") pod \"auto-csr-approver-29558018-kk95p\" (UID: \"5c586de6-9329-471d-b657-11668068ed0f\") " pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.438220 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bj2r\" (UniqueName: \"kubernetes.io/projected/5c586de6-9329-471d-b657-11668068ed0f-kube-api-access-8bj2r\") pod \"auto-csr-approver-29558018-kk95p\" (UID: \"5c586de6-9329-471d-b657-11668068ed0f\") " pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.464040 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bj2r\" (UniqueName: \"kubernetes.io/projected/5c586de6-9329-471d-b657-11668068ed0f-kube-api-access-8bj2r\") pod \"auto-csr-approver-29558018-kk95p\" (UID: \"5c586de6-9329-471d-b657-11668068ed0f\") " pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.486944 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:00 crc kubenswrapper[4843]: I0314 09:38:00.913517 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558018-kk95p"] Mar 14 09:38:01 crc kubenswrapper[4843]: I0314 09:38:01.476930 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558018-kk95p" event={"ID":"5c586de6-9329-471d-b657-11668068ed0f","Type":"ContainerStarted","Data":"01c5dad0ea2061816a133dcfa0ece549e211255c09684b1867e613f66073da2d"} Mar 14 09:38:02 crc kubenswrapper[4843]: I0314 09:38:02.487192 4843 generic.go:334] "Generic (PLEG): container finished" podID="5c586de6-9329-471d-b657-11668068ed0f" containerID="47eaa3b0d32c0b43186213676645af02b3675f4ad010ce9ecaa539a14e44caa3" exitCode=0 Mar 14 09:38:02 crc kubenswrapper[4843]: I0314 09:38:02.487283 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558018-kk95p" event={"ID":"5c586de6-9329-471d-b657-11668068ed0f","Type":"ContainerDied","Data":"47eaa3b0d32c0b43186213676645af02b3675f4ad010ce9ecaa539a14e44caa3"} Mar 14 09:38:03 crc kubenswrapper[4843]: I0314 09:38:03.796210 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:03 crc kubenswrapper[4843]: I0314 09:38:03.893930 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bj2r\" (UniqueName: \"kubernetes.io/projected/5c586de6-9329-471d-b657-11668068ed0f-kube-api-access-8bj2r\") pod \"5c586de6-9329-471d-b657-11668068ed0f\" (UID: \"5c586de6-9329-471d-b657-11668068ed0f\") " Mar 14 09:38:03 crc kubenswrapper[4843]: I0314 09:38:03.901616 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c586de6-9329-471d-b657-11668068ed0f-kube-api-access-8bj2r" (OuterVolumeSpecName: "kube-api-access-8bj2r") pod "5c586de6-9329-471d-b657-11668068ed0f" (UID: "5c586de6-9329-471d-b657-11668068ed0f"). InnerVolumeSpecName "kube-api-access-8bj2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:38:03 crc kubenswrapper[4843]: I0314 09:38:03.995375 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bj2r\" (UniqueName: \"kubernetes.io/projected/5c586de6-9329-471d-b657-11668068ed0f-kube-api-access-8bj2r\") on node \"crc\" DevicePath \"\"" Mar 14 09:38:04 crc kubenswrapper[4843]: I0314 09:38:04.508207 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558018-kk95p" event={"ID":"5c586de6-9329-471d-b657-11668068ed0f","Type":"ContainerDied","Data":"01c5dad0ea2061816a133dcfa0ece549e211255c09684b1867e613f66073da2d"} Mar 14 09:38:04 crc kubenswrapper[4843]: I0314 09:38:04.508249 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01c5dad0ea2061816a133dcfa0ece549e211255c09684b1867e613f66073da2d" Mar 14 09:38:04 crc kubenswrapper[4843]: I0314 09:38:04.508333 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558018-kk95p" Mar 14 09:38:04 crc kubenswrapper[4843]: I0314 09:38:04.876367 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558012-nrxkk"] Mar 14 09:38:04 crc kubenswrapper[4843]: I0314 09:38:04.883433 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558012-nrxkk"] Mar 14 09:38:05 crc kubenswrapper[4843]: I0314 09:38:05.348726 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74cd6ce3-6aba-48e9-a268-76deeffbadb5" path="/var/lib/kubelet/pods/74cd6ce3-6aba-48e9-a268-76deeffbadb5/volumes" Mar 14 09:38:08 crc kubenswrapper[4843]: I0314 09:38:08.339563 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:38:08 crc kubenswrapper[4843]: E0314 09:38:08.340073 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:38:21 crc kubenswrapper[4843]: I0314 09:38:21.339905 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:38:21 crc kubenswrapper[4843]: E0314 09:38:21.340855 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:38:32 crc kubenswrapper[4843]: I0314 09:38:32.341813 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:38:32 crc kubenswrapper[4843]: E0314 09:38:32.342586 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:38:47 crc kubenswrapper[4843]: I0314 09:38:47.338625 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:38:47 crc kubenswrapper[4843]: E0314 09:38:47.339410 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:38:48 crc kubenswrapper[4843]: I0314 09:38:48.768791 4843 scope.go:117] "RemoveContainer" containerID="627afe853906b53970a37d1adcd44701fb7f26abbf7086812f1278f110f77a7b" Mar 14 09:38:58 crc kubenswrapper[4843]: I0314 09:38:58.339383 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:38:58 crc kubenswrapper[4843]: E0314 09:38:58.340371 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:39:10 crc kubenswrapper[4843]: I0314 09:39:10.338996 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:39:10 crc kubenswrapper[4843]: E0314 09:39:10.339778 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:39:22 crc kubenswrapper[4843]: I0314 09:39:22.338680 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:39:22 crc kubenswrapper[4843]: E0314 09:39:22.339602 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:39:33 crc kubenswrapper[4843]: I0314 09:39:33.354806 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:39:33 crc kubenswrapper[4843]: E0314 09:39:33.355915 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:39:48 crc kubenswrapper[4843]: I0314 09:39:48.338752 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:39:48 crc kubenswrapper[4843]: E0314 09:39:48.339424 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:39:58 crc kubenswrapper[4843]: I0314 09:39:58.051584 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/root-account-create-update-mcvtm"] Mar 14 09:39:58 crc kubenswrapper[4843]: I0314 09:39:58.060697 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/root-account-create-update-mcvtm"] Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.035831 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc"] Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.046327 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-db-create-bjzn8"] Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.055595 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-fea7-account-create-update-ttmsc"] Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.062594 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-db-create-bjzn8"] Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.348932 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34d732b4-52bc-42fd-9bf1-23c248b749e6" path="/var/lib/kubelet/pods/34d732b4-52bc-42fd-9bf1-23c248b749e6/volumes" Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.349562 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="515aa0ff-df97-4090-9867-6ee2769d6750" path="/var/lib/kubelet/pods/515aa0ff-df97-4090-9867-6ee2769d6750/volumes" Mar 14 09:39:59 crc kubenswrapper[4843]: I0314 09:39:59.350335 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec047e38-cf88-4628-99ee-7aee2303e09f" path="/var/lib/kubelet/pods/ec047e38-cf88-4628-99ee-7aee2303e09f/volumes" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.153495 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558020-cz48j"] Mar 14 09:40:00 crc kubenswrapper[4843]: E0314 09:40:00.153850 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c586de6-9329-471d-b657-11668068ed0f" containerName="oc" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.153862 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c586de6-9329-471d-b657-11668068ed0f" containerName="oc" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.181198 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c586de6-9329-471d-b657-11668068ed0f" containerName="oc" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.182004 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.187207 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558020-cz48j"] Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.190858 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.191120 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.202101 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.272593 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml9sb\" (UniqueName: \"kubernetes.io/projected/060c6918-f03f-46f4-929d-b2b5c3d19369-kube-api-access-ml9sb\") pod \"auto-csr-approver-29558020-cz48j\" (UID: \"060c6918-f03f-46f4-929d-b2b5c3d19369\") " pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.373699 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml9sb\" (UniqueName: \"kubernetes.io/projected/060c6918-f03f-46f4-929d-b2b5c3d19369-kube-api-access-ml9sb\") pod \"auto-csr-approver-29558020-cz48j\" (UID: \"060c6918-f03f-46f4-929d-b2b5c3d19369\") " pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.390935 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml9sb\" (UniqueName: \"kubernetes.io/projected/060c6918-f03f-46f4-929d-b2b5c3d19369-kube-api-access-ml9sb\") pod \"auto-csr-approver-29558020-cz48j\" (UID: \"060c6918-f03f-46f4-929d-b2b5c3d19369\") " pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:00 crc kubenswrapper[4843]: I0314 09:40:00.529889 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:01 crc kubenswrapper[4843]: I0314 09:40:01.007009 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558020-cz48j"] Mar 14 09:40:01 crc kubenswrapper[4843]: I0314 09:40:01.019402 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:40:01 crc kubenswrapper[4843]: I0314 09:40:01.446117 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558020-cz48j" event={"ID":"060c6918-f03f-46f4-929d-b2b5c3d19369","Type":"ContainerStarted","Data":"b1cf9e508d3936df41647ac64d08c68c09ae12c369fdeffdb8a1c0dffddc4dbb"} Mar 14 09:40:03 crc kubenswrapper[4843]: I0314 09:40:03.347229 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:40:03 crc kubenswrapper[4843]: E0314 09:40:03.348055 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:40:03 crc kubenswrapper[4843]: I0314 09:40:03.467030 4843 generic.go:334] "Generic (PLEG): container finished" podID="060c6918-f03f-46f4-929d-b2b5c3d19369" containerID="be1e2a095ec3c07e68d7657f96f6998361e07f63972f4da47b84798d9f659778" exitCode=0 Mar 14 09:40:03 crc kubenswrapper[4843]: I0314 09:40:03.467122 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558020-cz48j" event={"ID":"060c6918-f03f-46f4-929d-b2b5c3d19369","Type":"ContainerDied","Data":"be1e2a095ec3c07e68d7657f96f6998361e07f63972f4da47b84798d9f659778"} Mar 14 09:40:04 crc kubenswrapper[4843]: I0314 09:40:04.798819 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:04 crc kubenswrapper[4843]: I0314 09:40:04.863303 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml9sb\" (UniqueName: \"kubernetes.io/projected/060c6918-f03f-46f4-929d-b2b5c3d19369-kube-api-access-ml9sb\") pod \"060c6918-f03f-46f4-929d-b2b5c3d19369\" (UID: \"060c6918-f03f-46f4-929d-b2b5c3d19369\") " Mar 14 09:40:04 crc kubenswrapper[4843]: I0314 09:40:04.869873 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060c6918-f03f-46f4-929d-b2b5c3d19369-kube-api-access-ml9sb" (OuterVolumeSpecName: "kube-api-access-ml9sb") pod "060c6918-f03f-46f4-929d-b2b5c3d19369" (UID: "060c6918-f03f-46f4-929d-b2b5c3d19369"). InnerVolumeSpecName "kube-api-access-ml9sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:40:04 crc kubenswrapper[4843]: I0314 09:40:04.965288 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml9sb\" (UniqueName: \"kubernetes.io/projected/060c6918-f03f-46f4-929d-b2b5c3d19369-kube-api-access-ml9sb\") on node \"crc\" DevicePath \"\"" Mar 14 09:40:05 crc kubenswrapper[4843]: I0314 09:40:05.486921 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558020-cz48j" event={"ID":"060c6918-f03f-46f4-929d-b2b5c3d19369","Type":"ContainerDied","Data":"b1cf9e508d3936df41647ac64d08c68c09ae12c369fdeffdb8a1c0dffddc4dbb"} Mar 14 09:40:05 crc kubenswrapper[4843]: I0314 09:40:05.487344 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1cf9e508d3936df41647ac64d08c68c09ae12c369fdeffdb8a1c0dffddc4dbb" Mar 14 09:40:05 crc kubenswrapper[4843]: I0314 09:40:05.486988 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558020-cz48j" Mar 14 09:40:05 crc kubenswrapper[4843]: I0314 09:40:05.850566 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558014-svhvt"] Mar 14 09:40:05 crc kubenswrapper[4843]: I0314 09:40:05.858436 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558014-svhvt"] Mar 14 09:40:07 crc kubenswrapper[4843]: I0314 09:40:07.349756 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2398db-31e4-49c5-bc8d-08157c7a3dbd" path="/var/lib/kubelet/pods/8d2398db-31e4-49c5-bc8d-08157c7a3dbd/volumes" Mar 14 09:40:18 crc kubenswrapper[4843]: I0314 09:40:18.338535 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:40:18 crc kubenswrapper[4843]: E0314 09:40:18.339299 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:40:19 crc kubenswrapper[4843]: I0314 09:40:19.047014 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kb65g"] Mar 14 09:40:19 crc kubenswrapper[4843]: I0314 09:40:19.072229 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-db-sync-kb65g"] Mar 14 09:40:19 crc kubenswrapper[4843]: I0314 09:40:19.350436 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d99019e-ec8b-4310-92c5-a764541ac1ce" path="/var/lib/kubelet/pods/8d99019e-ec8b-4310-92c5-a764541ac1ce/volumes" Mar 14 09:40:32 crc kubenswrapper[4843]: I0314 09:40:32.339245 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:40:32 crc kubenswrapper[4843]: E0314 09:40:32.339762 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:40:33 crc kubenswrapper[4843]: I0314 09:40:33.034829 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-cmr6j"] Mar 14 09:40:33 crc kubenswrapper[4843]: I0314 09:40:33.044326 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-cmr6j"] Mar 14 09:40:33 crc kubenswrapper[4843]: I0314 09:40:33.354563 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ca31b1f-982b-44fe-977d-6e2c9d2d007f" path="/var/lib/kubelet/pods/7ca31b1f-982b-44fe-977d-6e2c9d2d007f/volumes" Mar 14 09:40:47 crc kubenswrapper[4843]: I0314 09:40:47.339343 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:40:47 crc kubenswrapper[4843]: E0314 09:40:47.341129 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:40:48 crc kubenswrapper[4843]: I0314 09:40:48.904886 4843 scope.go:117] "RemoveContainer" containerID="984edb3276c5815657228407cea74e85d271d8b79c56c353c444ebabcc6a9664" Mar 14 09:40:48 crc kubenswrapper[4843]: I0314 09:40:48.953927 4843 scope.go:117] "RemoveContainer" containerID="c5fe0aad2cf927dc59be42fbab76aafa7361cdbaaf9b3954d808ac182459491e" Mar 14 09:40:48 crc kubenswrapper[4843]: I0314 09:40:48.975477 4843 scope.go:117] "RemoveContainer" containerID="c740779b5a9cfbed97f8ced9f50be3ca0d50483e67cad9c605deff98d43e1691" Mar 14 09:40:49 crc kubenswrapper[4843]: I0314 09:40:49.014214 4843 scope.go:117] "RemoveContainer" containerID="5b598fdfa71178c7ea0d960f8aaf659cf2e45cd1bac4489cdb34f277a86591fb" Mar 14 09:40:49 crc kubenswrapper[4843]: I0314 09:40:49.062529 4843 scope.go:117] "RemoveContainer" containerID="81131ff0a422c538240e57b47ebc08dc11b257ccbecb50723ef8483f8fa8e10b" Mar 14 09:40:49 crc kubenswrapper[4843]: I0314 09:40:49.086992 4843 scope.go:117] "RemoveContainer" containerID="3466ed6f81b7920e04e49be9c7db03edd9d3c4c4b7b28d209a022100a393cf79" Mar 14 09:40:58 crc kubenswrapper[4843]: I0314 09:40:58.338812 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:40:58 crc kubenswrapper[4843]: E0314 09:40:58.339490 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:41:11 crc kubenswrapper[4843]: I0314 09:41:11.339609 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:41:11 crc kubenswrapper[4843]: E0314 09:41:11.341875 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:41:23 crc kubenswrapper[4843]: I0314 09:41:23.346123 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:41:24 crc kubenswrapper[4843]: I0314 09:41:24.093436 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"1d86bf833c241f55a04a898386b9dd563f1910bdd2ccbe6124f8566a802dfff7"} Mar 14 09:41:29 crc kubenswrapper[4843]: I0314 09:41:29.042896 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-2pj9j"] Mar 14 09:41:29 crc kubenswrapper[4843]: I0314 09:41:29.050964 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-6f63-account-create-update-dx22q"] Mar 14 09:41:29 crc kubenswrapper[4843]: I0314 09:41:29.057774 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-2pj9j"] Mar 14 09:41:29 crc kubenswrapper[4843]: I0314 09:41:29.064614 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-6f63-account-create-update-dx22q"] Mar 14 09:41:29 crc kubenswrapper[4843]: I0314 09:41:29.350359 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4765db79-5f59-4bdf-a86e-4f69fa09b1c0" path="/var/lib/kubelet/pods/4765db79-5f59-4bdf-a86e-4f69fa09b1c0/volumes" Mar 14 09:41:29 crc kubenswrapper[4843]: I0314 09:41:29.351467 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5378344b-d874-4589-aab2-78186101e8d0" path="/var/lib/kubelet/pods/5378344b-d874-4589-aab2-78186101e8d0/volumes" Mar 14 09:41:49 crc kubenswrapper[4843]: I0314 09:41:49.237093 4843 scope.go:117] "RemoveContainer" containerID="b08e7f51d48295ca3bcd9303e4b7ecbe20e705e3bc7ea69f20fd2986cd82addf" Mar 14 09:41:49 crc kubenswrapper[4843]: I0314 09:41:49.264062 4843 scope.go:117] "RemoveContainer" containerID="903d9897aa84e59904dd95bc27b534b765bf5fe3a87107d2dbc1516d8b8cbdd7" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.140544 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558022-g74p7"] Mar 14 09:42:00 crc kubenswrapper[4843]: E0314 09:42:00.141396 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060c6918-f03f-46f4-929d-b2b5c3d19369" containerName="oc" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.141410 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="060c6918-f03f-46f4-929d-b2b5c3d19369" containerName="oc" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.141564 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="060c6918-f03f-46f4-929d-b2b5c3d19369" containerName="oc" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.142088 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.144114 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.144306 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.144306 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.150117 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558022-g74p7"] Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.264933 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8wrx\" (UniqueName: \"kubernetes.io/projected/3ed054ee-d91b-43df-ba81-b0abbe047bc1-kube-api-access-g8wrx\") pod \"auto-csr-approver-29558022-g74p7\" (UID: \"3ed054ee-d91b-43df-ba81-b0abbe047bc1\") " pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.367021 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8wrx\" (UniqueName: \"kubernetes.io/projected/3ed054ee-d91b-43df-ba81-b0abbe047bc1-kube-api-access-g8wrx\") pod \"auto-csr-approver-29558022-g74p7\" (UID: \"3ed054ee-d91b-43df-ba81-b0abbe047bc1\") " pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.386244 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8wrx\" (UniqueName: \"kubernetes.io/projected/3ed054ee-d91b-43df-ba81-b0abbe047bc1-kube-api-access-g8wrx\") pod \"auto-csr-approver-29558022-g74p7\" (UID: \"3ed054ee-d91b-43df-ba81-b0abbe047bc1\") " pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.463379 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:00 crc kubenswrapper[4843]: I0314 09:42:00.885429 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558022-g74p7"] Mar 14 09:42:01 crc kubenswrapper[4843]: I0314 09:42:01.421815 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558022-g74p7" event={"ID":"3ed054ee-d91b-43df-ba81-b0abbe047bc1","Type":"ContainerStarted","Data":"37548aba6584ed590c6bf0b5674deedd885c329f8cc9c060b559866907356e55"} Mar 14 09:42:02 crc kubenswrapper[4843]: I0314 09:42:02.431224 4843 generic.go:334] "Generic (PLEG): container finished" podID="3ed054ee-d91b-43df-ba81-b0abbe047bc1" containerID="3a56a1c9e4716e7c3635615ee804b0e8548d002a37dd3e6841aca2a0b44d9b9c" exitCode=0 Mar 14 09:42:02 crc kubenswrapper[4843]: I0314 09:42:02.431313 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558022-g74p7" event={"ID":"3ed054ee-d91b-43df-ba81-b0abbe047bc1","Type":"ContainerDied","Data":"3a56a1c9e4716e7c3635615ee804b0e8548d002a37dd3e6841aca2a0b44d9b9c"} Mar 14 09:42:03 crc kubenswrapper[4843]: I0314 09:42:03.808846 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:03 crc kubenswrapper[4843]: I0314 09:42:03.923250 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8wrx\" (UniqueName: \"kubernetes.io/projected/3ed054ee-d91b-43df-ba81-b0abbe047bc1-kube-api-access-g8wrx\") pod \"3ed054ee-d91b-43df-ba81-b0abbe047bc1\" (UID: \"3ed054ee-d91b-43df-ba81-b0abbe047bc1\") " Mar 14 09:42:03 crc kubenswrapper[4843]: I0314 09:42:03.929453 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ed054ee-d91b-43df-ba81-b0abbe047bc1-kube-api-access-g8wrx" (OuterVolumeSpecName: "kube-api-access-g8wrx") pod "3ed054ee-d91b-43df-ba81-b0abbe047bc1" (UID: "3ed054ee-d91b-43df-ba81-b0abbe047bc1"). InnerVolumeSpecName "kube-api-access-g8wrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:42:04 crc kubenswrapper[4843]: I0314 09:42:04.025353 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8wrx\" (UniqueName: \"kubernetes.io/projected/3ed054ee-d91b-43df-ba81-b0abbe047bc1-kube-api-access-g8wrx\") on node \"crc\" DevicePath \"\"" Mar 14 09:42:04 crc kubenswrapper[4843]: I0314 09:42:04.449209 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558022-g74p7" event={"ID":"3ed054ee-d91b-43df-ba81-b0abbe047bc1","Type":"ContainerDied","Data":"37548aba6584ed590c6bf0b5674deedd885c329f8cc9c060b559866907356e55"} Mar 14 09:42:04 crc kubenswrapper[4843]: I0314 09:42:04.449244 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558022-g74p7" Mar 14 09:42:04 crc kubenswrapper[4843]: I0314 09:42:04.449264 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37548aba6584ed590c6bf0b5674deedd885c329f8cc9c060b559866907356e55" Mar 14 09:42:04 crc kubenswrapper[4843]: I0314 09:42:04.873519 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558016-9r6w4"] Mar 14 09:42:04 crc kubenswrapper[4843]: I0314 09:42:04.881765 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558016-9r6w4"] Mar 14 09:42:05 crc kubenswrapper[4843]: I0314 09:42:05.348008 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43022031-bb5c-498e-8ddb-933b3058a45f" path="/var/lib/kubelet/pods/43022031-bb5c-498e-8ddb-933b3058a45f/volumes" Mar 14 09:42:07 crc kubenswrapper[4843]: I0314 09:42:07.028045 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sc292"] Mar 14 09:42:07 crc kubenswrapper[4843]: I0314 09:42:07.035292 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sc292"] Mar 14 09:42:07 crc kubenswrapper[4843]: I0314 09:42:07.348165 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3bee036-655a-484c-b818-12bbdb0ddfe5" path="/var/lib/kubelet/pods/a3bee036-655a-484c-b818-12bbdb0ddfe5/volumes" Mar 14 09:42:49 crc kubenswrapper[4843]: I0314 09:42:49.348322 4843 scope.go:117] "RemoveContainer" containerID="1fc7936c99da80ed1c59476ba191691097198e4678919c894a43b157662fb41e" Mar 14 09:42:49 crc kubenswrapper[4843]: I0314 09:42:49.388507 4843 scope.go:117] "RemoveContainer" containerID="db9356172cd17aa975a4dd73199e3644929a4395cd0dcfd07644f61db70898db" Mar 14 09:43:45 crc kubenswrapper[4843]: I0314 09:43:45.387219 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:43:45 crc kubenswrapper[4843]: I0314 09:43:45.387664 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.142161 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558024-mgk5q"] Mar 14 09:44:00 crc kubenswrapper[4843]: E0314 09:44:00.143031 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed054ee-d91b-43df-ba81-b0abbe047bc1" containerName="oc" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.143047 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed054ee-d91b-43df-ba81-b0abbe047bc1" containerName="oc" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.143254 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed054ee-d91b-43df-ba81-b0abbe047bc1" containerName="oc" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.143962 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.152117 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558024-mgk5q"] Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.153427 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.153439 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.153988 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.206504 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czfjf\" (UniqueName: \"kubernetes.io/projected/36fd9387-d70a-4b8d-9d0a-9f6b507432df-kube-api-access-czfjf\") pod \"auto-csr-approver-29558024-mgk5q\" (UID: \"36fd9387-d70a-4b8d-9d0a-9f6b507432df\") " pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.308646 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czfjf\" (UniqueName: \"kubernetes.io/projected/36fd9387-d70a-4b8d-9d0a-9f6b507432df-kube-api-access-czfjf\") pod \"auto-csr-approver-29558024-mgk5q\" (UID: \"36fd9387-d70a-4b8d-9d0a-9f6b507432df\") " pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.330841 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czfjf\" (UniqueName: \"kubernetes.io/projected/36fd9387-d70a-4b8d-9d0a-9f6b507432df-kube-api-access-czfjf\") pod \"auto-csr-approver-29558024-mgk5q\" (UID: \"36fd9387-d70a-4b8d-9d0a-9f6b507432df\") " pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.463187 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:00 crc kubenswrapper[4843]: I0314 09:44:00.927745 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558024-mgk5q"] Mar 14 09:44:01 crc kubenswrapper[4843]: I0314 09:44:01.632919 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" event={"ID":"36fd9387-d70a-4b8d-9d0a-9f6b507432df","Type":"ContainerStarted","Data":"d2ee9409726c7241c671566bda67d8ccb09fb004d446822b51fdd2dca71f25e2"} Mar 14 09:44:02 crc kubenswrapper[4843]: I0314 09:44:02.643645 4843 generic.go:334] "Generic (PLEG): container finished" podID="36fd9387-d70a-4b8d-9d0a-9f6b507432df" containerID="29cef80e012d63a5c0a221b4bf1dc5afb0078ca7589dbc0102184f7aaa699255" exitCode=0 Mar 14 09:44:02 crc kubenswrapper[4843]: I0314 09:44:02.643724 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" event={"ID":"36fd9387-d70a-4b8d-9d0a-9f6b507432df","Type":"ContainerDied","Data":"29cef80e012d63a5c0a221b4bf1dc5afb0078ca7589dbc0102184f7aaa699255"} Mar 14 09:44:03 crc kubenswrapper[4843]: I0314 09:44:03.951161 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:03 crc kubenswrapper[4843]: I0314 09:44:03.971063 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czfjf\" (UniqueName: \"kubernetes.io/projected/36fd9387-d70a-4b8d-9d0a-9f6b507432df-kube-api-access-czfjf\") pod \"36fd9387-d70a-4b8d-9d0a-9f6b507432df\" (UID: \"36fd9387-d70a-4b8d-9d0a-9f6b507432df\") " Mar 14 09:44:03 crc kubenswrapper[4843]: I0314 09:44:03.986634 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36fd9387-d70a-4b8d-9d0a-9f6b507432df-kube-api-access-czfjf" (OuterVolumeSpecName: "kube-api-access-czfjf") pod "36fd9387-d70a-4b8d-9d0a-9f6b507432df" (UID: "36fd9387-d70a-4b8d-9d0a-9f6b507432df"). InnerVolumeSpecName "kube-api-access-czfjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:44:04 crc kubenswrapper[4843]: I0314 09:44:04.073668 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czfjf\" (UniqueName: \"kubernetes.io/projected/36fd9387-d70a-4b8d-9d0a-9f6b507432df-kube-api-access-czfjf\") on node \"crc\" DevicePath \"\"" Mar 14 09:44:04 crc kubenswrapper[4843]: I0314 09:44:04.661054 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" event={"ID":"36fd9387-d70a-4b8d-9d0a-9f6b507432df","Type":"ContainerDied","Data":"d2ee9409726c7241c671566bda67d8ccb09fb004d446822b51fdd2dca71f25e2"} Mar 14 09:44:04 crc kubenswrapper[4843]: I0314 09:44:04.661093 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2ee9409726c7241c671566bda67d8ccb09fb004d446822b51fdd2dca71f25e2" Mar 14 09:44:04 crc kubenswrapper[4843]: I0314 09:44:04.661113 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558024-mgk5q" Mar 14 09:44:05 crc kubenswrapper[4843]: I0314 09:44:05.043049 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558018-kk95p"] Mar 14 09:44:05 crc kubenswrapper[4843]: I0314 09:44:05.049594 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558018-kk95p"] Mar 14 09:44:05 crc kubenswrapper[4843]: I0314 09:44:05.349211 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c586de6-9329-471d-b657-11668068ed0f" path="/var/lib/kubelet/pods/5c586de6-9329-471d-b657-11668068ed0f/volumes" Mar 14 09:44:15 crc kubenswrapper[4843]: I0314 09:44:15.387718 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:44:15 crc kubenswrapper[4843]: I0314 09:44:15.388322 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.413353 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mjv82"] Mar 14 09:44:23 crc kubenswrapper[4843]: E0314 09:44:23.414634 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36fd9387-d70a-4b8d-9d0a-9f6b507432df" containerName="oc" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.414664 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="36fd9387-d70a-4b8d-9d0a-9f6b507432df" containerName="oc" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.415008 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="36fd9387-d70a-4b8d-9d0a-9f6b507432df" containerName="oc" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.417415 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.431299 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mjv82"] Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.576525 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk2zd\" (UniqueName: \"kubernetes.io/projected/70e219e1-0adc-4368-b16a-5e5cd0561780-kube-api-access-kk2zd\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.576655 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-utilities\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.576679 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-catalog-content\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.677784 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk2zd\" (UniqueName: \"kubernetes.io/projected/70e219e1-0adc-4368-b16a-5e5cd0561780-kube-api-access-kk2zd\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.677893 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-utilities\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.677913 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-catalog-content\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.678330 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-catalog-content\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.678827 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-utilities\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.696493 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk2zd\" (UniqueName: \"kubernetes.io/projected/70e219e1-0adc-4368-b16a-5e5cd0561780-kube-api-access-kk2zd\") pod \"redhat-operators-mjv82\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:23 crc kubenswrapper[4843]: I0314 09:44:23.750666 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:24 crc kubenswrapper[4843]: I0314 09:44:24.188482 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mjv82"] Mar 14 09:44:24 crc kubenswrapper[4843]: I0314 09:44:24.822616 4843 generic.go:334] "Generic (PLEG): container finished" podID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerID="a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864" exitCode=0 Mar 14 09:44:24 crc kubenswrapper[4843]: I0314 09:44:24.822658 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjv82" event={"ID":"70e219e1-0adc-4368-b16a-5e5cd0561780","Type":"ContainerDied","Data":"a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864"} Mar 14 09:44:24 crc kubenswrapper[4843]: I0314 09:44:24.822684 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjv82" event={"ID":"70e219e1-0adc-4368-b16a-5e5cd0561780","Type":"ContainerStarted","Data":"c6d65ece744a8feaa1be9f2a2c1f2944779f494ee87dddade1a4af3917478911"} Mar 14 09:44:26 crc kubenswrapper[4843]: I0314 09:44:26.837542 4843 generic.go:334] "Generic (PLEG): container finished" podID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerID="1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c" exitCode=0 Mar 14 09:44:26 crc kubenswrapper[4843]: I0314 09:44:26.837576 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjv82" event={"ID":"70e219e1-0adc-4368-b16a-5e5cd0561780","Type":"ContainerDied","Data":"1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c"} Mar 14 09:44:27 crc kubenswrapper[4843]: I0314 09:44:27.847458 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjv82" event={"ID":"70e219e1-0adc-4368-b16a-5e5cd0561780","Type":"ContainerStarted","Data":"597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3"} Mar 14 09:44:27 crc kubenswrapper[4843]: I0314 09:44:27.873058 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mjv82" podStartSLOduration=2.424874907 podStartE2EDuration="4.873037814s" podCreationTimestamp="2026-03-14 09:44:23 +0000 UTC" firstStartedPulling="2026-03-14 09:44:24.824532538 +0000 UTC m=+2032.137143666" lastFinishedPulling="2026-03-14 09:44:27.272695445 +0000 UTC m=+2034.585306573" observedRunningTime="2026-03-14 09:44:27.86601519 +0000 UTC m=+2035.178626328" watchObservedRunningTime="2026-03-14 09:44:27.873037814 +0000 UTC m=+2035.185648942" Mar 14 09:44:33 crc kubenswrapper[4843]: I0314 09:44:33.751551 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:33 crc kubenswrapper[4843]: I0314 09:44:33.752502 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:33 crc kubenswrapper[4843]: I0314 09:44:33.803145 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:33 crc kubenswrapper[4843]: I0314 09:44:33.948658 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.387685 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mjv82"] Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.388392 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mjv82" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="registry-server" containerID="cri-o://597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3" gracePeriod=2 Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.788471 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.907848 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-utilities\") pod \"70e219e1-0adc-4368-b16a-5e5cd0561780\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.907959 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk2zd\" (UniqueName: \"kubernetes.io/projected/70e219e1-0adc-4368-b16a-5e5cd0561780-kube-api-access-kk2zd\") pod \"70e219e1-0adc-4368-b16a-5e5cd0561780\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.908108 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-catalog-content\") pod \"70e219e1-0adc-4368-b16a-5e5cd0561780\" (UID: \"70e219e1-0adc-4368-b16a-5e5cd0561780\") " Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.908655 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-utilities" (OuterVolumeSpecName: "utilities") pod "70e219e1-0adc-4368-b16a-5e5cd0561780" (UID: "70e219e1-0adc-4368-b16a-5e5cd0561780"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.916504 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70e219e1-0adc-4368-b16a-5e5cd0561780-kube-api-access-kk2zd" (OuterVolumeSpecName: "kube-api-access-kk2zd") pod "70e219e1-0adc-4368-b16a-5e5cd0561780" (UID: "70e219e1-0adc-4368-b16a-5e5cd0561780"). InnerVolumeSpecName "kube-api-access-kk2zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.922787 4843 generic.go:334] "Generic (PLEG): container finished" podID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerID="597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3" exitCode=0 Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.922826 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjv82" event={"ID":"70e219e1-0adc-4368-b16a-5e5cd0561780","Type":"ContainerDied","Data":"597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3"} Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.922852 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjv82" event={"ID":"70e219e1-0adc-4368-b16a-5e5cd0561780","Type":"ContainerDied","Data":"c6d65ece744a8feaa1be9f2a2c1f2944779f494ee87dddade1a4af3917478911"} Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.922868 4843 scope.go:117] "RemoveContainer" containerID="597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.922974 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjv82" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.955761 4843 scope.go:117] "RemoveContainer" containerID="1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c" Mar 14 09:44:37 crc kubenswrapper[4843]: I0314 09:44:37.975157 4843 scope.go:117] "RemoveContainer" containerID="a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.004418 4843 scope.go:117] "RemoveContainer" containerID="597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3" Mar 14 09:44:38 crc kubenswrapper[4843]: E0314 09:44:38.004845 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3\": container with ID starting with 597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3 not found: ID does not exist" containerID="597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.004887 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3"} err="failed to get container status \"597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3\": rpc error: code = NotFound desc = could not find container \"597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3\": container with ID starting with 597c826f21313662a619d6a627afda167aa16f944c28ef2926d3e1a3f5c786d3 not found: ID does not exist" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.004913 4843 scope.go:117] "RemoveContainer" containerID="1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c" Mar 14 09:44:38 crc kubenswrapper[4843]: E0314 09:44:38.005204 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c\": container with ID starting with 1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c not found: ID does not exist" containerID="1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.005254 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c"} err="failed to get container status \"1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c\": rpc error: code = NotFound desc = could not find container \"1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c\": container with ID starting with 1cbd0e8f4652b4768486c2e32bc0d2e27fc71f2d57f0b6ebdd0b046360b5960c not found: ID does not exist" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.005298 4843 scope.go:117] "RemoveContainer" containerID="a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864" Mar 14 09:44:38 crc kubenswrapper[4843]: E0314 09:44:38.005559 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864\": container with ID starting with a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864 not found: ID does not exist" containerID="a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.005582 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864"} err="failed to get container status \"a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864\": rpc error: code = NotFound desc = could not find container \"a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864\": container with ID starting with a707a83ace33992ef3496061e98849817f9483756012c2df434febef9570c864 not found: ID does not exist" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.009570 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.009593 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk2zd\" (UniqueName: \"kubernetes.io/projected/70e219e1-0adc-4368-b16a-5e5cd0561780-kube-api-access-kk2zd\") on node \"crc\" DevicePath \"\"" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.035185 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70e219e1-0adc-4368-b16a-5e5cd0561780" (UID: "70e219e1-0adc-4368-b16a-5e5cd0561780"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.110991 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e219e1-0adc-4368-b16a-5e5cd0561780-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.251007 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mjv82"] Mar 14 09:44:38 crc kubenswrapper[4843]: I0314 09:44:38.257548 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mjv82"] Mar 14 09:44:39 crc kubenswrapper[4843]: I0314 09:44:39.349993 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" path="/var/lib/kubelet/pods/70e219e1-0adc-4368-b16a-5e5cd0561780/volumes" Mar 14 09:44:45 crc kubenswrapper[4843]: I0314 09:44:45.387609 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:44:45 crc kubenswrapper[4843]: I0314 09:44:45.388082 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:44:45 crc kubenswrapper[4843]: I0314 09:44:45.388122 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:44:45 crc kubenswrapper[4843]: I0314 09:44:45.388751 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d86bf833c241f55a04a898386b9dd563f1910bdd2ccbe6124f8566a802dfff7"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:44:45 crc kubenswrapper[4843]: I0314 09:44:45.388802 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://1d86bf833c241f55a04a898386b9dd563f1910bdd2ccbe6124f8566a802dfff7" gracePeriod=600 Mar 14 09:44:46 crc kubenswrapper[4843]: I0314 09:44:46.013677 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="1d86bf833c241f55a04a898386b9dd563f1910bdd2ccbe6124f8566a802dfff7" exitCode=0 Mar 14 09:44:46 crc kubenswrapper[4843]: I0314 09:44:46.013785 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"1d86bf833c241f55a04a898386b9dd563f1910bdd2ccbe6124f8566a802dfff7"} Mar 14 09:44:46 crc kubenswrapper[4843]: I0314 09:44:46.014188 4843 scope.go:117] "RemoveContainer" containerID="0bba704465450a37098d299d90d210599bc51520cc074f7389ebbf11651b9bb0" Mar 14 09:44:46 crc kubenswrapper[4843]: I0314 09:44:46.014049 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba"} Mar 14 09:44:49 crc kubenswrapper[4843]: I0314 09:44:49.468059 4843 scope.go:117] "RemoveContainer" containerID="47eaa3b0d32c0b43186213676645af02b3675f4ad010ce9ecaa539a14e44caa3" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.142431 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz"] Mar 14 09:45:00 crc kubenswrapper[4843]: E0314 09:45:00.143255 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="extract-utilities" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.143284 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="extract-utilities" Mar 14 09:45:00 crc kubenswrapper[4843]: E0314 09:45:00.143294 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="registry-server" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.143300 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="registry-server" Mar 14 09:45:00 crc kubenswrapper[4843]: E0314 09:45:00.143312 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="extract-content" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.143320 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="extract-content" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.143475 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="70e219e1-0adc-4368-b16a-5e5cd0561780" containerName="registry-server" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.144041 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.146072 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.146546 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.156913 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz"] Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.315063 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b616c8e5-ade8-4e14-bf70-3d1887901ced-config-volume\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.315141 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8w5b\" (UniqueName: \"kubernetes.io/projected/b616c8e5-ade8-4e14-bf70-3d1887901ced-kube-api-access-b8w5b\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.315208 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b616c8e5-ade8-4e14-bf70-3d1887901ced-secret-volume\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.416538 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b616c8e5-ade8-4e14-bf70-3d1887901ced-config-volume\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.417670 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b616c8e5-ade8-4e14-bf70-3d1887901ced-config-volume\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.417860 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8w5b\" (UniqueName: \"kubernetes.io/projected/b616c8e5-ade8-4e14-bf70-3d1887901ced-kube-api-access-b8w5b\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.418733 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b616c8e5-ade8-4e14-bf70-3d1887901ced-secret-volume\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.428563 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b616c8e5-ade8-4e14-bf70-3d1887901ced-secret-volume\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.442723 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8w5b\" (UniqueName: \"kubernetes.io/projected/b616c8e5-ade8-4e14-bf70-3d1887901ced-kube-api-access-b8w5b\") pod \"collect-profiles-29558025-9wsdz\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:00 crc kubenswrapper[4843]: I0314 09:45:00.460727 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:01 crc kubenswrapper[4843]: I0314 09:45:01.171821 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz"] Mar 14 09:45:01 crc kubenswrapper[4843]: I0314 09:45:01.714719 4843 generic.go:334] "Generic (PLEG): container finished" podID="b616c8e5-ade8-4e14-bf70-3d1887901ced" containerID="2b26021568a49146ed2db8be7e77be78c4a01882d80c8f34784095bd99f22d38" exitCode=0 Mar 14 09:45:01 crc kubenswrapper[4843]: I0314 09:45:01.714765 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" event={"ID":"b616c8e5-ade8-4e14-bf70-3d1887901ced","Type":"ContainerDied","Data":"2b26021568a49146ed2db8be7e77be78c4a01882d80c8f34784095bd99f22d38"} Mar 14 09:45:01 crc kubenswrapper[4843]: I0314 09:45:01.714803 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" event={"ID":"b616c8e5-ade8-4e14-bf70-3d1887901ced","Type":"ContainerStarted","Data":"a27f2b412f45f2063e245dc568f15b9f81962d676e2c17dc042ccea68cfffb28"} Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.071033 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.159764 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b616c8e5-ade8-4e14-bf70-3d1887901ced-secret-volume\") pod \"b616c8e5-ade8-4e14-bf70-3d1887901ced\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.159820 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8w5b\" (UniqueName: \"kubernetes.io/projected/b616c8e5-ade8-4e14-bf70-3d1887901ced-kube-api-access-b8w5b\") pod \"b616c8e5-ade8-4e14-bf70-3d1887901ced\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.165140 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b616c8e5-ade8-4e14-bf70-3d1887901ced-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b616c8e5-ade8-4e14-bf70-3d1887901ced" (UID: "b616c8e5-ade8-4e14-bf70-3d1887901ced"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.165530 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b616c8e5-ade8-4e14-bf70-3d1887901ced-kube-api-access-b8w5b" (OuterVolumeSpecName: "kube-api-access-b8w5b") pod "b616c8e5-ade8-4e14-bf70-3d1887901ced" (UID: "b616c8e5-ade8-4e14-bf70-3d1887901ced"). InnerVolumeSpecName "kube-api-access-b8w5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.263155 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b616c8e5-ade8-4e14-bf70-3d1887901ced-config-volume\") pod \"b616c8e5-ade8-4e14-bf70-3d1887901ced\" (UID: \"b616c8e5-ade8-4e14-bf70-3d1887901ced\") " Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.264996 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b616c8e5-ade8-4e14-bf70-3d1887901ced-config-volume" (OuterVolumeSpecName: "config-volume") pod "b616c8e5-ade8-4e14-bf70-3d1887901ced" (UID: "b616c8e5-ade8-4e14-bf70-3d1887901ced"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.266672 4843 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b616c8e5-ade8-4e14-bf70-3d1887901ced-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.266699 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8w5b\" (UniqueName: \"kubernetes.io/projected/b616c8e5-ade8-4e14-bf70-3d1887901ced-kube-api-access-b8w5b\") on node \"crc\" DevicePath \"\"" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.266708 4843 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b616c8e5-ade8-4e14-bf70-3d1887901ced-config-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.734724 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" event={"ID":"b616c8e5-ade8-4e14-bf70-3d1887901ced","Type":"ContainerDied","Data":"a27f2b412f45f2063e245dc568f15b9f81962d676e2c17dc042ccea68cfffb28"} Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.735115 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a27f2b412f45f2063e245dc568f15b9f81962d676e2c17dc042ccea68cfffb28" Mar 14 09:45:03 crc kubenswrapper[4843]: I0314 09:45:03.734881 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558025-9wsdz" Mar 14 09:45:04 crc kubenswrapper[4843]: I0314 09:45:04.142758 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6"] Mar 14 09:45:04 crc kubenswrapper[4843]: I0314 09:45:04.150215 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557980-gk2g6"] Mar 14 09:45:05 crc kubenswrapper[4843]: I0314 09:45:05.348539 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35935053-e284-4215-b14c-efcb387201c4" path="/var/lib/kubelet/pods/35935053-e284-4215-b14c-efcb387201c4/volumes" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.000419 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mc2s9"] Mar 14 09:45:26 crc kubenswrapper[4843]: E0314 09:45:26.001478 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b616c8e5-ade8-4e14-bf70-3d1887901ced" containerName="collect-profiles" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.001495 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b616c8e5-ade8-4e14-bf70-3d1887901ced" containerName="collect-profiles" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.001707 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b616c8e5-ade8-4e14-bf70-3d1887901ced" containerName="collect-profiles" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.005576 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.017804 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mc2s9"] Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.153344 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-utilities\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.153426 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-catalog-content\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.153492 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcjkx\" (UniqueName: \"kubernetes.io/projected/41893d61-c19f-4e40-aa17-64d9ef9370ab-kube-api-access-lcjkx\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.254588 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcjkx\" (UniqueName: \"kubernetes.io/projected/41893d61-c19f-4e40-aa17-64d9ef9370ab-kube-api-access-lcjkx\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.254721 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-utilities\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.254771 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-catalog-content\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.255177 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-utilities\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.255306 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-catalog-content\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.279494 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcjkx\" (UniqueName: \"kubernetes.io/projected/41893d61-c19f-4e40-aa17-64d9ef9370ab-kube-api-access-lcjkx\") pod \"redhat-marketplace-mc2s9\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.327669 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.780565 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mc2s9"] Mar 14 09:45:26 crc kubenswrapper[4843]: I0314 09:45:26.915243 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mc2s9" event={"ID":"41893d61-c19f-4e40-aa17-64d9ef9370ab","Type":"ContainerStarted","Data":"3a67ebb7aa0377a9db9220905274ad476ebadcf1b21f8ea553d2efc13d2b6b93"} Mar 14 09:45:27 crc kubenswrapper[4843]: I0314 09:45:27.925641 4843 generic.go:334] "Generic (PLEG): container finished" podID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerID="e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2" exitCode=0 Mar 14 09:45:27 crc kubenswrapper[4843]: I0314 09:45:27.925711 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mc2s9" event={"ID":"41893d61-c19f-4e40-aa17-64d9ef9370ab","Type":"ContainerDied","Data":"e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2"} Mar 14 09:45:27 crc kubenswrapper[4843]: I0314 09:45:27.927926 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:45:28 crc kubenswrapper[4843]: I0314 09:45:28.940199 4843 generic.go:334] "Generic (PLEG): container finished" podID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerID="6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b" exitCode=0 Mar 14 09:45:28 crc kubenswrapper[4843]: I0314 09:45:28.940604 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mc2s9" event={"ID":"41893d61-c19f-4e40-aa17-64d9ef9370ab","Type":"ContainerDied","Data":"6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b"} Mar 14 09:45:29 crc kubenswrapper[4843]: I0314 09:45:29.953622 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mc2s9" event={"ID":"41893d61-c19f-4e40-aa17-64d9ef9370ab","Type":"ContainerStarted","Data":"ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb"} Mar 14 09:45:29 crc kubenswrapper[4843]: I0314 09:45:29.983801 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mc2s9" podStartSLOduration=3.578580611 podStartE2EDuration="4.983777144s" podCreationTimestamp="2026-03-14 09:45:25 +0000 UTC" firstStartedPulling="2026-03-14 09:45:27.927686687 +0000 UTC m=+2095.240297805" lastFinishedPulling="2026-03-14 09:45:29.33288321 +0000 UTC m=+2096.645494338" observedRunningTime="2026-03-14 09:45:29.975773935 +0000 UTC m=+2097.288385063" watchObservedRunningTime="2026-03-14 09:45:29.983777144 +0000 UTC m=+2097.296388272" Mar 14 09:45:36 crc kubenswrapper[4843]: I0314 09:45:36.328260 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:36 crc kubenswrapper[4843]: I0314 09:45:36.328866 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:36 crc kubenswrapper[4843]: I0314 09:45:36.375405 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:37 crc kubenswrapper[4843]: I0314 09:45:37.054095 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:39 crc kubenswrapper[4843]: I0314 09:45:39.990400 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mc2s9"] Mar 14 09:45:39 crc kubenswrapper[4843]: I0314 09:45:39.991193 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mc2s9" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="registry-server" containerID="cri-o://ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb" gracePeriod=2 Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.539396 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.614572 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcjkx\" (UniqueName: \"kubernetes.io/projected/41893d61-c19f-4e40-aa17-64d9ef9370ab-kube-api-access-lcjkx\") pod \"41893d61-c19f-4e40-aa17-64d9ef9370ab\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.614653 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-catalog-content\") pod \"41893d61-c19f-4e40-aa17-64d9ef9370ab\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.614773 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-utilities\") pod \"41893d61-c19f-4e40-aa17-64d9ef9370ab\" (UID: \"41893d61-c19f-4e40-aa17-64d9ef9370ab\") " Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.615903 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-utilities" (OuterVolumeSpecName: "utilities") pod "41893d61-c19f-4e40-aa17-64d9ef9370ab" (UID: "41893d61-c19f-4e40-aa17-64d9ef9370ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.620217 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41893d61-c19f-4e40-aa17-64d9ef9370ab-kube-api-access-lcjkx" (OuterVolumeSpecName: "kube-api-access-lcjkx") pod "41893d61-c19f-4e40-aa17-64d9ef9370ab" (UID: "41893d61-c19f-4e40-aa17-64d9ef9370ab"). InnerVolumeSpecName "kube-api-access-lcjkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.641897 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41893d61-c19f-4e40-aa17-64d9ef9370ab" (UID: "41893d61-c19f-4e40-aa17-64d9ef9370ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.717897 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.717980 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcjkx\" (UniqueName: \"kubernetes.io/projected/41893d61-c19f-4e40-aa17-64d9ef9370ab-kube-api-access-lcjkx\") on node \"crc\" DevicePath \"\"" Mar 14 09:45:40 crc kubenswrapper[4843]: I0314 09:45:40.718001 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41893d61-c19f-4e40-aa17-64d9ef9370ab-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.040920 4843 generic.go:334] "Generic (PLEG): container finished" podID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerID="ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb" exitCode=0 Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.040965 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mc2s9" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.040983 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mc2s9" event={"ID":"41893d61-c19f-4e40-aa17-64d9ef9370ab","Type":"ContainerDied","Data":"ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb"} Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.041080 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mc2s9" event={"ID":"41893d61-c19f-4e40-aa17-64d9ef9370ab","Type":"ContainerDied","Data":"3a67ebb7aa0377a9db9220905274ad476ebadcf1b21f8ea553d2efc13d2b6b93"} Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.041113 4843 scope.go:117] "RemoveContainer" containerID="ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.068988 4843 scope.go:117] "RemoveContainer" containerID="6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.072682 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mc2s9"] Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.080802 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mc2s9"] Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.097754 4843 scope.go:117] "RemoveContainer" containerID="e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.120486 4843 scope.go:117] "RemoveContainer" containerID="ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb" Mar 14 09:45:41 crc kubenswrapper[4843]: E0314 09:45:41.121165 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb\": container with ID starting with ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb not found: ID does not exist" containerID="ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.121209 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb"} err="failed to get container status \"ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb\": rpc error: code = NotFound desc = could not find container \"ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb\": container with ID starting with ef46cf2153e72a3a8c04ee1d9a70cf6a01e273017f15210f9e4c15406ab5bedb not found: ID does not exist" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.121239 4843 scope.go:117] "RemoveContainer" containerID="6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b" Mar 14 09:45:41 crc kubenswrapper[4843]: E0314 09:45:41.121831 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b\": container with ID starting with 6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b not found: ID does not exist" containerID="6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.121865 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b"} err="failed to get container status \"6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b\": rpc error: code = NotFound desc = could not find container \"6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b\": container with ID starting with 6c6599dc7241b473b3deb8ba87f68b7f3f6d58cdfa9fe24e2ffa8645a1998f1b not found: ID does not exist" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.121903 4843 scope.go:117] "RemoveContainer" containerID="e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2" Mar 14 09:45:41 crc kubenswrapper[4843]: E0314 09:45:41.122570 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2\": container with ID starting with e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2 not found: ID does not exist" containerID="e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.122706 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2"} err="failed to get container status \"e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2\": rpc error: code = NotFound desc = could not find container \"e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2\": container with ID starting with e2ecd3450fb6455feb463ce3aed6495ed5c368f476b6d43d5fbf9987487abec2 not found: ID does not exist" Mar 14 09:45:41 crc kubenswrapper[4843]: I0314 09:45:41.347766 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" path="/var/lib/kubelet/pods/41893d61-c19f-4e40-aa17-64d9ef9370ab/volumes" Mar 14 09:45:49 crc kubenswrapper[4843]: I0314 09:45:49.539180 4843 scope.go:117] "RemoveContainer" containerID="cb175251cbd5d50cd19772b7f16f9388f5affe3b8c575fcc41ed78e061ce11a4" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.147605 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558026-fpp82"] Mar 14 09:46:00 crc kubenswrapper[4843]: E0314 09:46:00.148441 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="registry-server" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.148452 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="registry-server" Mar 14 09:46:00 crc kubenswrapper[4843]: E0314 09:46:00.148474 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="extract-content" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.148481 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="extract-content" Mar 14 09:46:00 crc kubenswrapper[4843]: E0314 09:46:00.148489 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="extract-utilities" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.148495 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="extract-utilities" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.148655 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="41893d61-c19f-4e40-aa17-64d9ef9370ab" containerName="registry-server" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.149234 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.156574 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.157173 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.157250 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.161589 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558026-fpp82"] Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.247708 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49zvg\" (UniqueName: \"kubernetes.io/projected/44bd6132-462c-4296-9c39-c16070153e7c-kube-api-access-49zvg\") pod \"auto-csr-approver-29558026-fpp82\" (UID: \"44bd6132-462c-4296-9c39-c16070153e7c\") " pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.350356 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49zvg\" (UniqueName: \"kubernetes.io/projected/44bd6132-462c-4296-9c39-c16070153e7c-kube-api-access-49zvg\") pod \"auto-csr-approver-29558026-fpp82\" (UID: \"44bd6132-462c-4296-9c39-c16070153e7c\") " pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.368260 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49zvg\" (UniqueName: \"kubernetes.io/projected/44bd6132-462c-4296-9c39-c16070153e7c-kube-api-access-49zvg\") pod \"auto-csr-approver-29558026-fpp82\" (UID: \"44bd6132-462c-4296-9c39-c16070153e7c\") " pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.467889 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:00 crc kubenswrapper[4843]: I0314 09:46:00.917046 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558026-fpp82"] Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.201710 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558026-fpp82" event={"ID":"44bd6132-462c-4296-9c39-c16070153e7c","Type":"ContainerStarted","Data":"bb41ba8f6c535fcb9f5324428eca5fb82069c458a37baa683224d4fde249ded3"} Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.796102 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4rq2r"] Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.797749 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.811548 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4rq2r"] Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.878543 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-catalog-content\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.878737 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c578\" (UniqueName: \"kubernetes.io/projected/e31dff0c-800a-406a-96cb-576700604135-kube-api-access-9c578\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.878830 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-utilities\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.980098 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-catalog-content\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.980206 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c578\" (UniqueName: \"kubernetes.io/projected/e31dff0c-800a-406a-96cb-576700604135-kube-api-access-9c578\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.980283 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-utilities\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.980816 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-utilities\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:01 crc kubenswrapper[4843]: I0314 09:46:01.980840 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-catalog-content\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:02 crc kubenswrapper[4843]: I0314 09:46:02.012531 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c578\" (UniqueName: \"kubernetes.io/projected/e31dff0c-800a-406a-96cb-576700604135-kube-api-access-9c578\") pod \"certified-operators-4rq2r\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:02 crc kubenswrapper[4843]: I0314 09:46:02.122261 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:02 crc kubenswrapper[4843]: I0314 09:46:02.210110 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558026-fpp82" event={"ID":"44bd6132-462c-4296-9c39-c16070153e7c","Type":"ContainerStarted","Data":"2f1821f65759fdd2e3f3091ac7d5acfe8befd75222d5455e64a6d01613041443"} Mar 14 09:46:02 crc kubenswrapper[4843]: I0314 09:46:02.245094 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558026-fpp82" podStartSLOduration=1.325841023 podStartE2EDuration="2.245069456s" podCreationTimestamp="2026-03-14 09:46:00 +0000 UTC" firstStartedPulling="2026-03-14 09:46:00.91110523 +0000 UTC m=+2128.223716358" lastFinishedPulling="2026-03-14 09:46:01.830333663 +0000 UTC m=+2129.142944791" observedRunningTime="2026-03-14 09:46:02.232408102 +0000 UTC m=+2129.545019230" watchObservedRunningTime="2026-03-14 09:46:02.245069456 +0000 UTC m=+2129.557680584" Mar 14 09:46:02 crc kubenswrapper[4843]: I0314 09:46:02.622197 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4rq2r"] Mar 14 09:46:03 crc kubenswrapper[4843]: I0314 09:46:03.221563 4843 generic.go:334] "Generic (PLEG): container finished" podID="e31dff0c-800a-406a-96cb-576700604135" containerID="e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30" exitCode=0 Mar 14 09:46:03 crc kubenswrapper[4843]: I0314 09:46:03.221847 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerDied","Data":"e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30"} Mar 14 09:46:03 crc kubenswrapper[4843]: I0314 09:46:03.221871 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerStarted","Data":"3f68a707a8a638359276dd3867a76b5fc0294e8089866eeaeba1b56c7b73d3d5"} Mar 14 09:46:03 crc kubenswrapper[4843]: I0314 09:46:03.225778 4843 generic.go:334] "Generic (PLEG): container finished" podID="44bd6132-462c-4296-9c39-c16070153e7c" containerID="2f1821f65759fdd2e3f3091ac7d5acfe8befd75222d5455e64a6d01613041443" exitCode=0 Mar 14 09:46:03 crc kubenswrapper[4843]: I0314 09:46:03.225802 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558026-fpp82" event={"ID":"44bd6132-462c-4296-9c39-c16070153e7c","Type":"ContainerDied","Data":"2f1821f65759fdd2e3f3091ac7d5acfe8befd75222d5455e64a6d01613041443"} Mar 14 09:46:04 crc kubenswrapper[4843]: I0314 09:46:04.234512 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerStarted","Data":"99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50"} Mar 14 09:46:04 crc kubenswrapper[4843]: I0314 09:46:04.559420 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:04 crc kubenswrapper[4843]: I0314 09:46:04.627613 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49zvg\" (UniqueName: \"kubernetes.io/projected/44bd6132-462c-4296-9c39-c16070153e7c-kube-api-access-49zvg\") pod \"44bd6132-462c-4296-9c39-c16070153e7c\" (UID: \"44bd6132-462c-4296-9c39-c16070153e7c\") " Mar 14 09:46:04 crc kubenswrapper[4843]: I0314 09:46:04.633013 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44bd6132-462c-4296-9c39-c16070153e7c-kube-api-access-49zvg" (OuterVolumeSpecName: "kube-api-access-49zvg") pod "44bd6132-462c-4296-9c39-c16070153e7c" (UID: "44bd6132-462c-4296-9c39-c16070153e7c"). InnerVolumeSpecName "kube-api-access-49zvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:46:04 crc kubenswrapper[4843]: I0314 09:46:04.729681 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49zvg\" (UniqueName: \"kubernetes.io/projected/44bd6132-462c-4296-9c39-c16070153e7c-kube-api-access-49zvg\") on node \"crc\" DevicePath \"\"" Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.245061 4843 generic.go:334] "Generic (PLEG): container finished" podID="e31dff0c-800a-406a-96cb-576700604135" containerID="99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50" exitCode=0 Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.245129 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerDied","Data":"99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50"} Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.249186 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558026-fpp82" event={"ID":"44bd6132-462c-4296-9c39-c16070153e7c","Type":"ContainerDied","Data":"bb41ba8f6c535fcb9f5324428eca5fb82069c458a37baa683224d4fde249ded3"} Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.249225 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb41ba8f6c535fcb9f5324428eca5fb82069c458a37baa683224d4fde249ded3" Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.249226 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558026-fpp82" Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.296981 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558020-cz48j"] Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.304143 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558020-cz48j"] Mar 14 09:46:05 crc kubenswrapper[4843]: I0314 09:46:05.348466 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="060c6918-f03f-46f4-929d-b2b5c3d19369" path="/var/lib/kubelet/pods/060c6918-f03f-46f4-929d-b2b5c3d19369/volumes" Mar 14 09:46:06 crc kubenswrapper[4843]: I0314 09:46:06.271202 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerStarted","Data":"e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537"} Mar 14 09:46:06 crc kubenswrapper[4843]: I0314 09:46:06.301538 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4rq2r" podStartSLOduration=2.846571071 podStartE2EDuration="5.301521327s" podCreationTimestamp="2026-03-14 09:46:01 +0000 UTC" firstStartedPulling="2026-03-14 09:46:03.223864497 +0000 UTC m=+2130.536475625" lastFinishedPulling="2026-03-14 09:46:05.678814743 +0000 UTC m=+2132.991425881" observedRunningTime="2026-03-14 09:46:06.300602174 +0000 UTC m=+2133.613213312" watchObservedRunningTime="2026-03-14 09:46:06.301521327 +0000 UTC m=+2133.614132455" Mar 14 09:46:12 crc kubenswrapper[4843]: I0314 09:46:12.122854 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:12 crc kubenswrapper[4843]: I0314 09:46:12.124085 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:12 crc kubenswrapper[4843]: I0314 09:46:12.173898 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:12 crc kubenswrapper[4843]: I0314 09:46:12.350220 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:15 crc kubenswrapper[4843]: I0314 09:46:15.790924 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4rq2r"] Mar 14 09:46:15 crc kubenswrapper[4843]: I0314 09:46:15.791507 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4rq2r" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="registry-server" containerID="cri-o://e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537" gracePeriod=2 Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.282587 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.318919 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-catalog-content\") pod \"e31dff0c-800a-406a-96cb-576700604135\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.318993 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c578\" (UniqueName: \"kubernetes.io/projected/e31dff0c-800a-406a-96cb-576700604135-kube-api-access-9c578\") pod \"e31dff0c-800a-406a-96cb-576700604135\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.319034 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-utilities\") pod \"e31dff0c-800a-406a-96cb-576700604135\" (UID: \"e31dff0c-800a-406a-96cb-576700604135\") " Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.320580 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-utilities" (OuterVolumeSpecName: "utilities") pod "e31dff0c-800a-406a-96cb-576700604135" (UID: "e31dff0c-800a-406a-96cb-576700604135"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.328546 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31dff0c-800a-406a-96cb-576700604135-kube-api-access-9c578" (OuterVolumeSpecName: "kube-api-access-9c578") pod "e31dff0c-800a-406a-96cb-576700604135" (UID: "e31dff0c-800a-406a-96cb-576700604135"). InnerVolumeSpecName "kube-api-access-9c578". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.363731 4843 generic.go:334] "Generic (PLEG): container finished" podID="e31dff0c-800a-406a-96cb-576700604135" containerID="e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537" exitCode=0 Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.363777 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerDied","Data":"e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537"} Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.363808 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rq2r" event={"ID":"e31dff0c-800a-406a-96cb-576700604135","Type":"ContainerDied","Data":"3f68a707a8a638359276dd3867a76b5fc0294e8089866eeaeba1b56c7b73d3d5"} Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.363828 4843 scope.go:117] "RemoveContainer" containerID="e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.363826 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rq2r" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.376405 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e31dff0c-800a-406a-96cb-576700604135" (UID: "e31dff0c-800a-406a-96cb-576700604135"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.390413 4843 scope.go:117] "RemoveContainer" containerID="99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.408485 4843 scope.go:117] "RemoveContainer" containerID="e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.422659 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.422696 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c578\" (UniqueName: \"kubernetes.io/projected/e31dff0c-800a-406a-96cb-576700604135-kube-api-access-9c578\") on node \"crc\" DevicePath \"\"" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.422711 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31dff0c-800a-406a-96cb-576700604135-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.448211 4843 scope.go:117] "RemoveContainer" containerID="e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537" Mar 14 09:46:16 crc kubenswrapper[4843]: E0314 09:46:16.448950 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537\": container with ID starting with e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537 not found: ID does not exist" containerID="e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.449006 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537"} err="failed to get container status \"e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537\": rpc error: code = NotFound desc = could not find container \"e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537\": container with ID starting with e8058981820ace61c98bef3b28beb23c5b382fe67a026550361ef42c1fc37537 not found: ID does not exist" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.449035 4843 scope.go:117] "RemoveContainer" containerID="99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50" Mar 14 09:46:16 crc kubenswrapper[4843]: E0314 09:46:16.449474 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50\": container with ID starting with 99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50 not found: ID does not exist" containerID="99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.449499 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50"} err="failed to get container status \"99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50\": rpc error: code = NotFound desc = could not find container \"99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50\": container with ID starting with 99fa69b7d743ff0fbb1fbfe23808dfc9291ded2308c1de8d3ab405bf7476cd50 not found: ID does not exist" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.449514 4843 scope.go:117] "RemoveContainer" containerID="e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30" Mar 14 09:46:16 crc kubenswrapper[4843]: E0314 09:46:16.449913 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30\": container with ID starting with e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30 not found: ID does not exist" containerID="e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.449980 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30"} err="failed to get container status \"e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30\": rpc error: code = NotFound desc = could not find container \"e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30\": container with ID starting with e98a919c24ed93a71bd7bd5d75f8f19d25cea100fc8d7ef51abb02ff16e9cf30 not found: ID does not exist" Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.694695 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4rq2r"] Mar 14 09:46:16 crc kubenswrapper[4843]: I0314 09:46:16.704624 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4rq2r"] Mar 14 09:46:17 crc kubenswrapper[4843]: I0314 09:46:17.368547 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31dff0c-800a-406a-96cb-576700604135" path="/var/lib/kubelet/pods/e31dff0c-800a-406a-96cb-576700604135/volumes" Mar 14 09:46:45 crc kubenswrapper[4843]: I0314 09:46:45.387206 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:46:45 crc kubenswrapper[4843]: I0314 09:46:45.388110 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:46:49 crc kubenswrapper[4843]: I0314 09:46:49.618559 4843 scope.go:117] "RemoveContainer" containerID="be1e2a095ec3c07e68d7657f96f6998361e07f63972f4da47b84798d9f659778" Mar 14 09:47:15 crc kubenswrapper[4843]: I0314 09:47:15.387304 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:47:15 crc kubenswrapper[4843]: I0314 09:47:15.387916 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.359159 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher6f63-account-delete-pjfxg"] Mar 14 09:47:22 crc kubenswrapper[4843]: E0314 09:47:22.360104 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="extract-content" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.360124 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="extract-content" Mar 14 09:47:22 crc kubenswrapper[4843]: E0314 09:47:22.360152 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="registry-server" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.360163 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="registry-server" Mar 14 09:47:22 crc kubenswrapper[4843]: E0314 09:47:22.360182 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="extract-utilities" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.360190 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="extract-utilities" Mar 14 09:47:22 crc kubenswrapper[4843]: E0314 09:47:22.360210 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44bd6132-462c-4296-9c39-c16070153e7c" containerName="oc" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.360217 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="44bd6132-462c-4296-9c39-c16070153e7c" containerName="oc" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.360436 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31dff0c-800a-406a-96cb-576700604135" containerName="registry-server" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.360463 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="44bd6132-462c-4296-9c39-c16070153e7c" containerName="oc" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.361210 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.388824 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher6f63-account-delete-pjfxg"] Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.455325 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.460399 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="80eda2d0-78f4-4bac-b183-5ecf731131bb" containerName="watcher-decision-engine" containerID="cri-o://6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45" gracePeriod=30 Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.466897 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.467122 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-kuttl-api-log" containerID="cri-o://7768f365244cf747cf507e0a18759142f1e50773fa5d722a07446ad397f247b2" gracePeriod=30 Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.467254 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-api" containerID="cri-o://f531fb376f64a529edacba1abe9989f5e77a7bf1713f5faf46b594293d49955d" gracePeriod=30 Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.515503 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.515695 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" containerName="watcher-applier" containerID="cri-o://0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" gracePeriod=30 Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.528515 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4msn4\" (UniqueName: \"kubernetes.io/projected/1da671bd-c309-4405-ae5f-0c1ac1be129d-kube-api-access-4msn4\") pod \"watcher6f63-account-delete-pjfxg\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.528587 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da671bd-c309-4405-ae5f-0c1ac1be129d-operator-scripts\") pod \"watcher6f63-account-delete-pjfxg\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.629717 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4msn4\" (UniqueName: \"kubernetes.io/projected/1da671bd-c309-4405-ae5f-0c1ac1be129d-kube-api-access-4msn4\") pod \"watcher6f63-account-delete-pjfxg\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.629796 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da671bd-c309-4405-ae5f-0c1ac1be129d-operator-scripts\") pod \"watcher6f63-account-delete-pjfxg\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.630555 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da671bd-c309-4405-ae5f-0c1ac1be129d-operator-scripts\") pod \"watcher6f63-account-delete-pjfxg\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.650945 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4msn4\" (UniqueName: \"kubernetes.io/projected/1da671bd-c309-4405-ae5f-0c1ac1be129d-kube-api-access-4msn4\") pod \"watcher6f63-account-delete-pjfxg\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.733381 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.970706 4843 generic.go:334] "Generic (PLEG): container finished" podID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerID="7768f365244cf747cf507e0a18759142f1e50773fa5d722a07446ad397f247b2" exitCode=143 Mar 14 09:47:22 crc kubenswrapper[4843]: I0314 09:47:22.970751 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"33a65ce6-9c33-447c-b19a-17ab9bdadbd4","Type":"ContainerDied","Data":"7768f365244cf747cf507e0a18759142f1e50773fa5d722a07446ad397f247b2"} Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.244532 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher6f63-account-delete-pjfxg"] Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.583341 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.140:9322/\": dial tcp 10.217.0.140:9322: connect: connection refused" Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.583354 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.140:9322/\": dial tcp 10.217.0.140:9322: connect: connection refused" Mar 14 09:47:23 crc kubenswrapper[4843]: E0314 09:47:23.693040 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:47:23 crc kubenswrapper[4843]: E0314 09:47:23.695981 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:47:23 crc kubenswrapper[4843]: E0314 09:47:23.697334 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:47:23 crc kubenswrapper[4843]: E0314 09:47:23.697387 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" containerName="watcher-applier" Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.987649 4843 generic.go:334] "Generic (PLEG): container finished" podID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerID="f531fb376f64a529edacba1abe9989f5e77a7bf1713f5faf46b594293d49955d" exitCode=0 Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.987701 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"33a65ce6-9c33-447c-b19a-17ab9bdadbd4","Type":"ContainerDied","Data":"f531fb376f64a529edacba1abe9989f5e77a7bf1713f5faf46b594293d49955d"} Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.989790 4843 generic.go:334] "Generic (PLEG): container finished" podID="1da671bd-c309-4405-ae5f-0c1ac1be129d" containerID="2a985a0a7d0935235b6fc177d8b636f30d439f1568a8d03efafeb33918bae25c" exitCode=0 Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.989823 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" event={"ID":"1da671bd-c309-4405-ae5f-0c1ac1be129d","Type":"ContainerDied","Data":"2a985a0a7d0935235b6fc177d8b636f30d439f1568a8d03efafeb33918bae25c"} Mar 14 09:47:23 crc kubenswrapper[4843]: I0314 09:47:23.989848 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" event={"ID":"1da671bd-c309-4405-ae5f-0c1ac1be129d","Type":"ContainerStarted","Data":"f5876315882304b3f78423daaea234d4ed34b5395476ee8beff29353e7068c3f"} Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.341231 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.459182 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-custom-prometheus-ca\") pod \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.459243 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-config-data\") pod \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.459285 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-logs\") pod \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.459435 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbqsw\" (UniqueName: \"kubernetes.io/projected/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-kube-api-access-fbqsw\") pod \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.459458 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-combined-ca-bundle\") pod \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\" (UID: \"33a65ce6-9c33-447c-b19a-17ab9bdadbd4\") " Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.461316 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-logs" (OuterVolumeSpecName: "logs") pod "33a65ce6-9c33-447c-b19a-17ab9bdadbd4" (UID: "33a65ce6-9c33-447c-b19a-17ab9bdadbd4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.469503 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-kube-api-access-fbqsw" (OuterVolumeSpecName: "kube-api-access-fbqsw") pod "33a65ce6-9c33-447c-b19a-17ab9bdadbd4" (UID: "33a65ce6-9c33-447c-b19a-17ab9bdadbd4"). InnerVolumeSpecName "kube-api-access-fbqsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.485997 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "33a65ce6-9c33-447c-b19a-17ab9bdadbd4" (UID: "33a65ce6-9c33-447c-b19a-17ab9bdadbd4"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.489313 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33a65ce6-9c33-447c-b19a-17ab9bdadbd4" (UID: "33a65ce6-9c33-447c-b19a-17ab9bdadbd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.505676 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-config-data" (OuterVolumeSpecName: "config-data") pod "33a65ce6-9c33-447c-b19a-17ab9bdadbd4" (UID: "33a65ce6-9c33-447c-b19a-17ab9bdadbd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.561640 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbqsw\" (UniqueName: \"kubernetes.io/projected/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-kube-api-access-fbqsw\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.561685 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.561707 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.561720 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:24 crc kubenswrapper[4843]: I0314 09:47:24.561732 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a65ce6-9c33-447c-b19a-17ab9bdadbd4-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.000832 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"33a65ce6-9c33-447c-b19a-17ab9bdadbd4","Type":"ContainerDied","Data":"45c0c669520d9f636e1b1ec8a84d7b6489cf12a7e23f96393e0cf05ca2331444"} Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.001495 4843 scope.go:117] "RemoveContainer" containerID="f531fb376f64a529edacba1abe9989f5e77a7bf1713f5faf46b594293d49955d" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.000855 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.031204 4843 scope.go:117] "RemoveContainer" containerID="7768f365244cf747cf507e0a18759142f1e50773fa5d722a07446ad397f247b2" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.037446 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.045668 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.201336 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.201604 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-central-agent" containerID="cri-o://c0d8665e6c2da59a74f6081aa691475086398513074aff57bae2acbeed63b983" gracePeriod=30 Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.201652 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-notification-agent" containerID="cri-o://84d2de8f9f78a9774751d0ae39ceb02c52c7688df02840a185f9247080f0b12f" gracePeriod=30 Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.201684 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="sg-core" containerID="cri-o://d45adcb661ce46daaed762166c2c44da177aee39ee9ad8d74f77764714d1c8e2" gracePeriod=30 Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.201889 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="proxy-httpd" containerID="cri-o://4e77f996f65b0aeddd27d3c8d93056e06c4ba0e348778c82695b0e0ea8c48eca" gracePeriod=30 Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.352618 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" path="/var/lib/kubelet/pods/33a65ce6-9c33-447c-b19a-17ab9bdadbd4/volumes" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.353247 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.476152 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da671bd-c309-4405-ae5f-0c1ac1be129d-operator-scripts\") pod \"1da671bd-c309-4405-ae5f-0c1ac1be129d\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.476339 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4msn4\" (UniqueName: \"kubernetes.io/projected/1da671bd-c309-4405-ae5f-0c1ac1be129d-kube-api-access-4msn4\") pod \"1da671bd-c309-4405-ae5f-0c1ac1be129d\" (UID: \"1da671bd-c309-4405-ae5f-0c1ac1be129d\") " Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.477403 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da671bd-c309-4405-ae5f-0c1ac1be129d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1da671bd-c309-4405-ae5f-0c1ac1be129d" (UID: "1da671bd-c309-4405-ae5f-0c1ac1be129d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.482513 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da671bd-c309-4405-ae5f-0c1ac1be129d-kube-api-access-4msn4" (OuterVolumeSpecName: "kube-api-access-4msn4") pod "1da671bd-c309-4405-ae5f-0c1ac1be129d" (UID: "1da671bd-c309-4405-ae5f-0c1ac1be129d"). InnerVolumeSpecName "kube-api-access-4msn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.577670 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da671bd-c309-4405-ae5f-0c1ac1be129d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:25 crc kubenswrapper[4843]: I0314 09:47:25.577707 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4msn4\" (UniqueName: \"kubernetes.io/projected/1da671bd-c309-4405-ae5f-0c1ac1be129d-kube-api-access-4msn4\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034078 4843 generic.go:334] "Generic (PLEG): container finished" podID="18cb7825-ada4-4690-a848-9e8379bb7237" containerID="4e77f996f65b0aeddd27d3c8d93056e06c4ba0e348778c82695b0e0ea8c48eca" exitCode=0 Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034363 4843 generic.go:334] "Generic (PLEG): container finished" podID="18cb7825-ada4-4690-a848-9e8379bb7237" containerID="d45adcb661ce46daaed762166c2c44da177aee39ee9ad8d74f77764714d1c8e2" exitCode=2 Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034379 4843 generic.go:334] "Generic (PLEG): container finished" podID="18cb7825-ada4-4690-a848-9e8379bb7237" containerID="84d2de8f9f78a9774751d0ae39ceb02c52c7688df02840a185f9247080f0b12f" exitCode=0 Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034389 4843 generic.go:334] "Generic (PLEG): container finished" podID="18cb7825-ada4-4690-a848-9e8379bb7237" containerID="c0d8665e6c2da59a74f6081aa691475086398513074aff57bae2acbeed63b983" exitCode=0 Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034138 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerDied","Data":"4e77f996f65b0aeddd27d3c8d93056e06c4ba0e348778c82695b0e0ea8c48eca"} Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034467 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerDied","Data":"d45adcb661ce46daaed762166c2c44da177aee39ee9ad8d74f77764714d1c8e2"} Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034486 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerDied","Data":"84d2de8f9f78a9774751d0ae39ceb02c52c7688df02840a185f9247080f0b12f"} Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.034503 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerDied","Data":"c0d8665e6c2da59a74f6081aa691475086398513074aff57bae2acbeed63b983"} Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.037072 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" event={"ID":"1da671bd-c309-4405-ae5f-0c1ac1be129d","Type":"ContainerDied","Data":"f5876315882304b3f78423daaea234d4ed34b5395476ee8beff29353e7068c3f"} Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.037111 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5876315882304b3f78423daaea234d4ed34b5395476ee8beff29353e7068c3f" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.037191 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher6f63-account-delete-pjfxg" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.219727 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.389747 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-ceilometer-tls-certs\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.389819 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdzzt\" (UniqueName: \"kubernetes.io/projected/18cb7825-ada4-4690-a848-9e8379bb7237-kube-api-access-rdzzt\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.389851 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-sg-core-conf-yaml\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.389884 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-combined-ca-bundle\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.389923 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-log-httpd\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.390013 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-config-data\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.390047 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-scripts\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.390120 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-run-httpd\") pod \"18cb7825-ada4-4690-a848-9e8379bb7237\" (UID: \"18cb7825-ada4-4690-a848-9e8379bb7237\") " Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.390758 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.390804 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.397417 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-scripts" (OuterVolumeSpecName: "scripts") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.397438 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18cb7825-ada4-4690-a848-9e8379bb7237-kube-api-access-rdzzt" (OuterVolumeSpecName: "kube-api-access-rdzzt") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "kube-api-access-rdzzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.430667 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.437849 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.471432 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.494288 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.494324 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.494336 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.494349 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdzzt\" (UniqueName: \"kubernetes.io/projected/18cb7825-ada4-4690-a848-9e8379bb7237-kube-api-access-rdzzt\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.494399 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.494962 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.495677 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18cb7825-ada4-4690-a848-9e8379bb7237-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.500381 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-config-data" (OuterVolumeSpecName: "config-data") pod "18cb7825-ada4-4690-a848-9e8379bb7237" (UID: "18cb7825-ada4-4690-a848-9e8379bb7237"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.597431 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18cb7825-ada4-4690-a848-9e8379bb7237-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:26 crc kubenswrapper[4843]: I0314 09:47:26.996738 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.052685 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"18cb7825-ada4-4690-a848-9e8379bb7237","Type":"ContainerDied","Data":"1d9cb778a5c6018784cd969ccf1e54bffc2d001ec2cdd2c538b4138790bcd1c4"} Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.052735 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.052751 4843 scope.go:117] "RemoveContainer" containerID="4e77f996f65b0aeddd27d3c8d93056e06c4ba0e348778c82695b0e0ea8c48eca" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.057499 4843 generic.go:334] "Generic (PLEG): container finished" podID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" exitCode=0 Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.057535 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b7e25e6d-3253-4ac0-a420-3452f4f8275f","Type":"ContainerDied","Data":"0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a"} Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.057564 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b7e25e6d-3253-4ac0-a420-3452f4f8275f","Type":"ContainerDied","Data":"f871e7e63af74bc13a158d279cefcd719c29af4053d04db452b6dc4967fd8e00"} Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.057573 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.084114 4843 scope.go:117] "RemoveContainer" containerID="d45adcb661ce46daaed762166c2c44da177aee39ee9ad8d74f77764714d1c8e2" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.104521 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.109904 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4b4q\" (UniqueName: \"kubernetes.io/projected/b7e25e6d-3253-4ac0-a420-3452f4f8275f-kube-api-access-p4b4q\") pod \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.109964 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7e25e6d-3253-4ac0-a420-3452f4f8275f-logs\") pod \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.110094 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-config-data\") pod \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.110130 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-combined-ca-bundle\") pod \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\" (UID: \"b7e25e6d-3253-4ac0-a420-3452f4f8275f\") " Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.113356 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e25e6d-3253-4ac0-a420-3452f4f8275f-logs" (OuterVolumeSpecName: "logs") pod "b7e25e6d-3253-4ac0-a420-3452f4f8275f" (UID: "b7e25e6d-3253-4ac0-a420-3452f4f8275f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.132345 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e25e6d-3253-4ac0-a420-3452f4f8275f-kube-api-access-p4b4q" (OuterVolumeSpecName: "kube-api-access-p4b4q") pod "b7e25e6d-3253-4ac0-a420-3452f4f8275f" (UID: "b7e25e6d-3253-4ac0-a420-3452f4f8275f"). InnerVolumeSpecName "kube-api-access-p4b4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.135242 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.149102 4843 scope.go:117] "RemoveContainer" containerID="84d2de8f9f78a9774751d0ae39ceb02c52c7688df02840a185f9247080f0b12f" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.160535 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.160932 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" containerName="watcher-applier" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.160951 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" containerName="watcher-applier" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.160963 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da671bd-c309-4405-ae5f-0c1ac1be129d" containerName="mariadb-account-delete" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.160971 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da671bd-c309-4405-ae5f-0c1ac1be129d" containerName="mariadb-account-delete" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.160984 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-notification-agent" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.160992 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-notification-agent" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.161003 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-kuttl-api-log" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161011 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-kuttl-api-log" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.161023 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-api" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161030 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-api" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.161045 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="sg-core" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161052 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="sg-core" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.161072 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-central-agent" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161079 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-central-agent" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.161092 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="proxy-httpd" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161099 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="proxy-httpd" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161286 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="sg-core" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161300 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-api" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161314 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-central-agent" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161331 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="proxy-httpd" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161340 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" containerName="watcher-applier" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161353 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a65ce6-9c33-447c-b19a-17ab9bdadbd4" containerName="watcher-kuttl-api-log" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161363 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" containerName="ceilometer-notification-agent" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.161377 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da671bd-c309-4405-ae5f-0c1ac1be129d" containerName="mariadb-account-delete" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.175921 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.196716 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.196890 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.197041 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.205477 4843 scope.go:117] "RemoveContainer" containerID="c0d8665e6c2da59a74f6081aa691475086398513074aff57bae2acbeed63b983" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.212451 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4b4q\" (UniqueName: \"kubernetes.io/projected/b7e25e6d-3253-4ac0-a420-3452f4f8275f-kube-api-access-p4b4q\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.212488 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7e25e6d-3253-4ac0-a420-3452f4f8275f-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.237611 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.246529 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7e25e6d-3253-4ac0-a420-3452f4f8275f" (UID: "b7e25e6d-3253-4ac0-a420-3452f4f8275f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.249254 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-config-data" (OuterVolumeSpecName: "config-data") pod "b7e25e6d-3253-4ac0-a420-3452f4f8275f" (UID: "b7e25e6d-3253-4ac0-a420-3452f4f8275f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.262651 4843 scope.go:117] "RemoveContainer" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316077 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316178 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316204 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-run-httpd\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316227 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-scripts\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316305 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316324 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qrwn\" (UniqueName: \"kubernetes.io/projected/0881ee30-9845-468c-8f39-581f429e437d-kube-api-access-4qrwn\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316341 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-config-data\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316362 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-log-httpd\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316412 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.316422 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e25e6d-3253-4ac0-a420-3452f4f8275f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.357715 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18cb7825-ada4-4690-a848-9e8379bb7237" path="/var/lib/kubelet/pods/18cb7825-ada4-4690-a848-9e8379bb7237/volumes" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.359674 4843 scope.go:117] "RemoveContainer" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" Mar 14 09:47:27 crc kubenswrapper[4843]: E0314 09:47:27.364537 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a\": container with ID starting with 0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a not found: ID does not exist" containerID="0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.364593 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a"} err="failed to get container status \"0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a\": rpc error: code = NotFound desc = could not find container \"0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a\": container with ID starting with 0eed3d866858d59bfd33418b85bec2f76ad7cd0663bb37c02ecd2e6084bb103a not found: ID does not exist" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.416150 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417492 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417610 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417659 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-run-httpd\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417692 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-scripts\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417744 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417769 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qrwn\" (UniqueName: \"kubernetes.io/projected/0881ee30-9845-468c-8f39-581f429e437d-kube-api-access-4qrwn\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417792 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-config-data\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.417822 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-log-httpd\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.418376 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-log-httpd\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.426189 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.430031 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-run-httpd\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.436357 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.437938 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.438236 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-config-data\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.439393 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher6f63-account-delete-pjfxg"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.439554 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.440175 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-scripts\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.448696 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qrwn\" (UniqueName: \"kubernetes.io/projected/0881ee30-9845-468c-8f39-581f429e437d-kube-api-access-4qrwn\") pod \"ceilometer-0\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.459721 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher6f63-account-delete-pjfxg"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.596575 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-b745-account-create-update-l5lcz"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.597907 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.599428 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.610638 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-vrl6s"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.611868 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.621598 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vrl6s"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.634796 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-b745-account-create-update-l5lcz"] Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.661178 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.722926 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsn98\" (UniqueName: \"kubernetes.io/projected/7f774281-e77b-4fcd-a025-1888c65e2c33-kube-api-access-vsn98\") pod \"watcher-b745-account-create-update-l5lcz\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.723267 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwkgf\" (UniqueName: \"kubernetes.io/projected/b2040650-3240-48f6-a544-4041d9230929-kube-api-access-mwkgf\") pod \"watcher-db-create-vrl6s\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.723326 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f774281-e77b-4fcd-a025-1888c65e2c33-operator-scripts\") pod \"watcher-b745-account-create-update-l5lcz\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.723375 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2040650-3240-48f6-a544-4041d9230929-operator-scripts\") pod \"watcher-db-create-vrl6s\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.824685 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsn98\" (UniqueName: \"kubernetes.io/projected/7f774281-e77b-4fcd-a025-1888c65e2c33-kube-api-access-vsn98\") pod \"watcher-b745-account-create-update-l5lcz\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.824737 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkgf\" (UniqueName: \"kubernetes.io/projected/b2040650-3240-48f6-a544-4041d9230929-kube-api-access-mwkgf\") pod \"watcher-db-create-vrl6s\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.824767 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f774281-e77b-4fcd-a025-1888c65e2c33-operator-scripts\") pod \"watcher-b745-account-create-update-l5lcz\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.824803 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2040650-3240-48f6-a544-4041d9230929-operator-scripts\") pod \"watcher-db-create-vrl6s\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.825614 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f774281-e77b-4fcd-a025-1888c65e2c33-operator-scripts\") pod \"watcher-b745-account-create-update-l5lcz\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.825877 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2040650-3240-48f6-a544-4041d9230929-operator-scripts\") pod \"watcher-db-create-vrl6s\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.845566 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwkgf\" (UniqueName: \"kubernetes.io/projected/b2040650-3240-48f6-a544-4041d9230929-kube-api-access-mwkgf\") pod \"watcher-db-create-vrl6s\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.849367 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsn98\" (UniqueName: \"kubernetes.io/projected/7f774281-e77b-4fcd-a025-1888c65e2c33-kube-api-access-vsn98\") pod \"watcher-b745-account-create-update-l5lcz\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.923453 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:27 crc kubenswrapper[4843]: I0314 09:47:27.929585 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.154895 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:28 crc kubenswrapper[4843]: W0314 09:47:28.164780 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0881ee30_9845_468c_8f39_581f429e437d.slice/crio-50b9debb09719ffd1a50fd05cea3b9c3d70018862d4b35a76ebe2bee463e495e WatchSource:0}: Error finding container 50b9debb09719ffd1a50fd05cea3b9c3d70018862d4b35a76ebe2bee463e495e: Status 404 returned error can't find the container with id 50b9debb09719ffd1a50fd05cea3b9c3d70018862d4b35a76ebe2bee463e495e Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.396197 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vrl6s"] Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.468622 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-b745-account-create-update-l5lcz"] Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.682913 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.842524 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psf2n\" (UniqueName: \"kubernetes.io/projected/80eda2d0-78f4-4bac-b183-5ecf731131bb-kube-api-access-psf2n\") pod \"80eda2d0-78f4-4bac-b183-5ecf731131bb\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.842610 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-custom-prometheus-ca\") pod \"80eda2d0-78f4-4bac-b183-5ecf731131bb\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.842645 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-config-data\") pod \"80eda2d0-78f4-4bac-b183-5ecf731131bb\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.842722 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80eda2d0-78f4-4bac-b183-5ecf731131bb-logs\") pod \"80eda2d0-78f4-4bac-b183-5ecf731131bb\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.842817 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-combined-ca-bundle\") pod \"80eda2d0-78f4-4bac-b183-5ecf731131bb\" (UID: \"80eda2d0-78f4-4bac-b183-5ecf731131bb\") " Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.843144 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80eda2d0-78f4-4bac-b183-5ecf731131bb-logs" (OuterVolumeSpecName: "logs") pod "80eda2d0-78f4-4bac-b183-5ecf731131bb" (UID: "80eda2d0-78f4-4bac-b183-5ecf731131bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.847876 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80eda2d0-78f4-4bac-b183-5ecf731131bb-kube-api-access-psf2n" (OuterVolumeSpecName: "kube-api-access-psf2n") pod "80eda2d0-78f4-4bac-b183-5ecf731131bb" (UID: "80eda2d0-78f4-4bac-b183-5ecf731131bb"). InnerVolumeSpecName "kube-api-access-psf2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.867478 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80eda2d0-78f4-4bac-b183-5ecf731131bb" (UID: "80eda2d0-78f4-4bac-b183-5ecf731131bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.875948 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "80eda2d0-78f4-4bac-b183-5ecf731131bb" (UID: "80eda2d0-78f4-4bac-b183-5ecf731131bb"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.896256 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-config-data" (OuterVolumeSpecName: "config-data") pod "80eda2d0-78f4-4bac-b183-5ecf731131bb" (UID: "80eda2d0-78f4-4bac-b183-5ecf731131bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.944501 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80eda2d0-78f4-4bac-b183-5ecf731131bb-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.944770 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.944782 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psf2n\" (UniqueName: \"kubernetes.io/projected/80eda2d0-78f4-4bac-b183-5ecf731131bb-kube-api-access-psf2n\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.944791 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:28 crc kubenswrapper[4843]: I0314 09:47:28.944800 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80eda2d0-78f4-4bac-b183-5ecf731131bb-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.097441 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerStarted","Data":"50b9debb09719ffd1a50fd05cea3b9c3d70018862d4b35a76ebe2bee463e495e"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.099215 4843 generic.go:334] "Generic (PLEG): container finished" podID="80eda2d0-78f4-4bac-b183-5ecf731131bb" containerID="6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45" exitCode=0 Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.099289 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"80eda2d0-78f4-4bac-b183-5ecf731131bb","Type":"ContainerDied","Data":"6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.099300 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.099313 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"80eda2d0-78f4-4bac-b183-5ecf731131bb","Type":"ContainerDied","Data":"d43df2b06c99a71f87f6edce513b583fb32028f5b14cc17d60ea11cc40cb5428"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.099328 4843 scope.go:117] "RemoveContainer" containerID="6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.101201 4843 generic.go:334] "Generic (PLEG): container finished" podID="7f774281-e77b-4fcd-a025-1888c65e2c33" containerID="741291ee1cf83addf57cdd65999f6388854c5c20b1503a2e0c56d4de682f450c" exitCode=0 Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.101282 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" event={"ID":"7f774281-e77b-4fcd-a025-1888c65e2c33","Type":"ContainerDied","Data":"741291ee1cf83addf57cdd65999f6388854c5c20b1503a2e0c56d4de682f450c"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.101318 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" event={"ID":"7f774281-e77b-4fcd-a025-1888c65e2c33","Type":"ContainerStarted","Data":"95593620ea2949a02b8f0736336b9514ea1fe55c3470c076069ac94955a8d5dd"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.103333 4843 generic.go:334] "Generic (PLEG): container finished" podID="b2040650-3240-48f6-a544-4041d9230929" containerID="163b59a8f00ed4d144a7d383210261c8b1ddbc5e0ba3122195dce7602d3e6de2" exitCode=0 Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.103362 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vrl6s" event={"ID":"b2040650-3240-48f6-a544-4041d9230929","Type":"ContainerDied","Data":"163b59a8f00ed4d144a7d383210261c8b1ddbc5e0ba3122195dce7602d3e6de2"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.103380 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vrl6s" event={"ID":"b2040650-3240-48f6-a544-4041d9230929","Type":"ContainerStarted","Data":"fefcd94f6de146549e4eba968262a8058440e11bfb5fe715d6391e7931ba0fb6"} Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.125245 4843 scope.go:117] "RemoveContainer" containerID="6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45" Mar 14 09:47:29 crc kubenswrapper[4843]: E0314 09:47:29.126754 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45\": container with ID starting with 6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45 not found: ID does not exist" containerID="6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.126792 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45"} err="failed to get container status \"6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45\": rpc error: code = NotFound desc = could not find container \"6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45\": container with ID starting with 6d236f5b86df3d3a92578eb28f9a35e175345af8d31a41194bd4620e65f08f45 not found: ID does not exist" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.155201 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.162634 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.347989 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da671bd-c309-4405-ae5f-0c1ac1be129d" path="/var/lib/kubelet/pods/1da671bd-c309-4405-ae5f-0c1ac1be129d/volumes" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.348533 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80eda2d0-78f4-4bac-b183-5ecf731131bb" path="/var/lib/kubelet/pods/80eda2d0-78f4-4bac-b183-5ecf731131bb/volumes" Mar 14 09:47:29 crc kubenswrapper[4843]: I0314 09:47:29.349030 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e25e6d-3253-4ac0-a420-3452f4f8275f" path="/var/lib/kubelet/pods/b7e25e6d-3253-4ac0-a420-3452f4f8275f/volumes" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.113554 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerStarted","Data":"421d18b2d5b4387373e03df685083b5e4f778c57ff9740e308f1a0faaddb43aa"} Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.114045 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerStarted","Data":"730b02f435cd77dcc4150583e231f1d1082e0efe72e629aa440264b8dacdddbf"} Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.569122 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.594504 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.684049 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsn98\" (UniqueName: \"kubernetes.io/projected/7f774281-e77b-4fcd-a025-1888c65e2c33-kube-api-access-vsn98\") pod \"7f774281-e77b-4fcd-a025-1888c65e2c33\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.684093 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f774281-e77b-4fcd-a025-1888c65e2c33-operator-scripts\") pod \"7f774281-e77b-4fcd-a025-1888c65e2c33\" (UID: \"7f774281-e77b-4fcd-a025-1888c65e2c33\") " Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.684187 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwkgf\" (UniqueName: \"kubernetes.io/projected/b2040650-3240-48f6-a544-4041d9230929-kube-api-access-mwkgf\") pod \"b2040650-3240-48f6-a544-4041d9230929\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.684218 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2040650-3240-48f6-a544-4041d9230929-operator-scripts\") pod \"b2040650-3240-48f6-a544-4041d9230929\" (UID: \"b2040650-3240-48f6-a544-4041d9230929\") " Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.684708 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f774281-e77b-4fcd-a025-1888c65e2c33-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f774281-e77b-4fcd-a025-1888c65e2c33" (UID: "7f774281-e77b-4fcd-a025-1888c65e2c33"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.684993 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2040650-3240-48f6-a544-4041d9230929-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2040650-3240-48f6-a544-4041d9230929" (UID: "b2040650-3240-48f6-a544-4041d9230929"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.690507 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2040650-3240-48f6-a544-4041d9230929-kube-api-access-mwkgf" (OuterVolumeSpecName: "kube-api-access-mwkgf") pod "b2040650-3240-48f6-a544-4041d9230929" (UID: "b2040650-3240-48f6-a544-4041d9230929"). InnerVolumeSpecName "kube-api-access-mwkgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.705533 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f774281-e77b-4fcd-a025-1888c65e2c33-kube-api-access-vsn98" (OuterVolumeSpecName: "kube-api-access-vsn98") pod "7f774281-e77b-4fcd-a025-1888c65e2c33" (UID: "7f774281-e77b-4fcd-a025-1888c65e2c33"). InnerVolumeSpecName "kube-api-access-vsn98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.786242 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsn98\" (UniqueName: \"kubernetes.io/projected/7f774281-e77b-4fcd-a025-1888c65e2c33-kube-api-access-vsn98\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.786360 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f774281-e77b-4fcd-a025-1888c65e2c33-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.786372 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwkgf\" (UniqueName: \"kubernetes.io/projected/b2040650-3240-48f6-a544-4041d9230929-kube-api-access-mwkgf\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:30 crc kubenswrapper[4843]: I0314 09:47:30.786381 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2040650-3240-48f6-a544-4041d9230929-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.123147 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.123159 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-b745-account-create-update-l5lcz" event={"ID":"7f774281-e77b-4fcd-a025-1888c65e2c33","Type":"ContainerDied","Data":"95593620ea2949a02b8f0736336b9514ea1fe55c3470c076069ac94955a8d5dd"} Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.123883 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95593620ea2949a02b8f0736336b9514ea1fe55c3470c076069ac94955a8d5dd" Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.124955 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-vrl6s" event={"ID":"b2040650-3240-48f6-a544-4041d9230929","Type":"ContainerDied","Data":"fefcd94f6de146549e4eba968262a8058440e11bfb5fe715d6391e7931ba0fb6"} Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.124988 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-vrl6s" Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.124993 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fefcd94f6de146549e4eba968262a8058440e11bfb5fe715d6391e7931ba0fb6" Mar 14 09:47:31 crc kubenswrapper[4843]: I0314 09:47:31.126875 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerStarted","Data":"bdb38ab86dde4597b71f9e291c0b65ef2455275a8379e5c0dfd81aa3bd49a4fe"} Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.945994 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5"] Mar 14 09:47:32 crc kubenswrapper[4843]: E0314 09:47:32.946506 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f774281-e77b-4fcd-a025-1888c65e2c33" containerName="mariadb-account-create-update" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.946517 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f774281-e77b-4fcd-a025-1888c65e2c33" containerName="mariadb-account-create-update" Mar 14 09:47:32 crc kubenswrapper[4843]: E0314 09:47:32.946536 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80eda2d0-78f4-4bac-b183-5ecf731131bb" containerName="watcher-decision-engine" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.946541 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="80eda2d0-78f4-4bac-b183-5ecf731131bb" containerName="watcher-decision-engine" Mar 14 09:47:32 crc kubenswrapper[4843]: E0314 09:47:32.946561 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2040650-3240-48f6-a544-4041d9230929" containerName="mariadb-database-create" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.946567 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2040650-3240-48f6-a544-4041d9230929" containerName="mariadb-database-create" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.946709 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="80eda2d0-78f4-4bac-b183-5ecf731131bb" containerName="watcher-decision-engine" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.946722 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2040650-3240-48f6-a544-4041d9230929" containerName="mariadb-database-create" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.946730 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f774281-e77b-4fcd-a025-1888c65e2c33" containerName="mariadb-account-create-update" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.947244 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.951863 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-gggnj" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.957216 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:47:32 crc kubenswrapper[4843]: I0314 09:47:32.959616 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5"] Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.122563 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-db-sync-config-data\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.122681 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.122702 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m7tn\" (UniqueName: \"kubernetes.io/projected/3f4ffebb-2312-4d73-8409-8cfb8ac69041-kube-api-access-9m7tn\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.122721 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-config-data\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.143285 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerStarted","Data":"6c7c6e0ba8581c41e5aba5676bf98300c6d7e98edc891045f5e1ad64d90e2106"} Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.144248 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.177462 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.9557496749999999 podStartE2EDuration="6.177436739s" podCreationTimestamp="2026-03-14 09:47:27 +0000 UTC" firstStartedPulling="2026-03-14 09:47:28.167223839 +0000 UTC m=+2215.479834967" lastFinishedPulling="2026-03-14 09:47:32.388910903 +0000 UTC m=+2219.701522031" observedRunningTime="2026-03-14 09:47:33.170139355 +0000 UTC m=+2220.482750483" watchObservedRunningTime="2026-03-14 09:47:33.177436739 +0000 UTC m=+2220.490047877" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.224039 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-db-sync-config-data\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.224148 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.224171 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m7tn\" (UniqueName: \"kubernetes.io/projected/3f4ffebb-2312-4d73-8409-8cfb8ac69041-kube-api-access-9m7tn\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.224193 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-config-data\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.229104 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-config-data\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.229396 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-db-sync-config-data\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.230056 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.242041 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m7tn\" (UniqueName: \"kubernetes.io/projected/3f4ffebb-2312-4d73-8409-8cfb8ac69041-kube-api-access-9m7tn\") pod \"watcher-kuttl-db-sync-lbwt5\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.276687 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:33 crc kubenswrapper[4843]: I0314 09:47:33.735067 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5"] Mar 14 09:47:33 crc kubenswrapper[4843]: W0314 09:47:33.739232 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f4ffebb_2312_4d73_8409_8cfb8ac69041.slice/crio-c3ed6e92aa6ae924aca30711aa7fab88ef028d11f2874015e38ed2acd6f7f2c4 WatchSource:0}: Error finding container c3ed6e92aa6ae924aca30711aa7fab88ef028d11f2874015e38ed2acd6f7f2c4: Status 404 returned error can't find the container with id c3ed6e92aa6ae924aca30711aa7fab88ef028d11f2874015e38ed2acd6f7f2c4 Mar 14 09:47:34 crc kubenswrapper[4843]: I0314 09:47:34.152763 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" event={"ID":"3f4ffebb-2312-4d73-8409-8cfb8ac69041","Type":"ContainerStarted","Data":"8124c7ce40ba05f5165d0a40f17fcda6f92cdc98efb114f09bfe50443426a48a"} Mar 14 09:47:34 crc kubenswrapper[4843]: I0314 09:47:34.153034 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" event={"ID":"3f4ffebb-2312-4d73-8409-8cfb8ac69041","Type":"ContainerStarted","Data":"c3ed6e92aa6ae924aca30711aa7fab88ef028d11f2874015e38ed2acd6f7f2c4"} Mar 14 09:47:34 crc kubenswrapper[4843]: I0314 09:47:34.180091 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" podStartSLOduration=2.180074859 podStartE2EDuration="2.180074859s" podCreationTimestamp="2026-03-14 09:47:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:47:34.177245771 +0000 UTC m=+2221.489856899" watchObservedRunningTime="2026-03-14 09:47:34.180074859 +0000 UTC m=+2221.492685987" Mar 14 09:47:37 crc kubenswrapper[4843]: I0314 09:47:37.181196 4843 generic.go:334] "Generic (PLEG): container finished" podID="3f4ffebb-2312-4d73-8409-8cfb8ac69041" containerID="8124c7ce40ba05f5165d0a40f17fcda6f92cdc98efb114f09bfe50443426a48a" exitCode=0 Mar 14 09:47:37 crc kubenswrapper[4843]: I0314 09:47:37.181370 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" event={"ID":"3f4ffebb-2312-4d73-8409-8cfb8ac69041","Type":"ContainerDied","Data":"8124c7ce40ba05f5165d0a40f17fcda6f92cdc98efb114f09bfe50443426a48a"} Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.480776 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.622692 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m7tn\" (UniqueName: \"kubernetes.io/projected/3f4ffebb-2312-4d73-8409-8cfb8ac69041-kube-api-access-9m7tn\") pod \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.622830 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-config-data\") pod \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.622894 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-db-sync-config-data\") pod \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.622914 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-combined-ca-bundle\") pod \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\" (UID: \"3f4ffebb-2312-4d73-8409-8cfb8ac69041\") " Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.628190 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3f4ffebb-2312-4d73-8409-8cfb8ac69041" (UID: "3f4ffebb-2312-4d73-8409-8cfb8ac69041"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.630851 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f4ffebb-2312-4d73-8409-8cfb8ac69041-kube-api-access-9m7tn" (OuterVolumeSpecName: "kube-api-access-9m7tn") pod "3f4ffebb-2312-4d73-8409-8cfb8ac69041" (UID: "3f4ffebb-2312-4d73-8409-8cfb8ac69041"). InnerVolumeSpecName "kube-api-access-9m7tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.652514 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f4ffebb-2312-4d73-8409-8cfb8ac69041" (UID: "3f4ffebb-2312-4d73-8409-8cfb8ac69041"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.665297 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-config-data" (OuterVolumeSpecName: "config-data") pod "3f4ffebb-2312-4d73-8409-8cfb8ac69041" (UID: "3f4ffebb-2312-4d73-8409-8cfb8ac69041"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.725195 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m7tn\" (UniqueName: \"kubernetes.io/projected/3f4ffebb-2312-4d73-8409-8cfb8ac69041-kube-api-access-9m7tn\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.725242 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.725257 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:38 crc kubenswrapper[4843]: I0314 09:47:38.725285 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f4ffebb-2312-4d73-8409-8cfb8ac69041-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.203064 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" event={"ID":"3f4ffebb-2312-4d73-8409-8cfb8ac69041","Type":"ContainerDied","Data":"c3ed6e92aa6ae924aca30711aa7fab88ef028d11f2874015e38ed2acd6f7f2c4"} Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.203120 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3ed6e92aa6ae924aca30711aa7fab88ef028d11f2874015e38ed2acd6f7f2c4" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.203181 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.546003 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:39 crc kubenswrapper[4843]: E0314 09:47:39.546431 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4ffebb-2312-4d73-8409-8cfb8ac69041" containerName="watcher-kuttl-db-sync" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.546449 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4ffebb-2312-4d73-8409-8cfb8ac69041" containerName="watcher-kuttl-db-sync" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.546703 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f4ffebb-2312-4d73-8409-8cfb8ac69041" containerName="watcher-kuttl-db-sync" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.547379 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.549859 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-gggnj" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.555470 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.560822 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.562853 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.568362 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.570944 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.588674 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.630497 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.631791 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.637595 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.639749 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2e103b-d279-4a6f-bf32-985a0389730c-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.639811 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhlch\" (UniqueName: \"kubernetes.io/projected/1b2e103b-d279-4a6f-bf32-985a0389730c-kube-api-access-jhlch\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.639876 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.639899 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.640065 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rv66\" (UniqueName: \"kubernetes.io/projected/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-kube-api-access-4rv66\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.640135 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.640162 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.640389 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.640439 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-logs\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.665468 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742366 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742472 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rv66\" (UniqueName: \"kubernetes.io/projected/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-kube-api-access-4rv66\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742506 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742522 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742550 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742570 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-logs\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742598 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e6b57e-42c7-4e03-8f1e-813a66d93644-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742638 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742680 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742710 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4hg\" (UniqueName: \"kubernetes.io/projected/62e6b57e-42c7-4e03-8f1e-813a66d93644-kube-api-access-qd4hg\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742736 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2e103b-d279-4a6f-bf32-985a0389730c-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742764 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhlch\" (UniqueName: \"kubernetes.io/projected/1b2e103b-d279-4a6f-bf32-985a0389730c-kube-api-access-jhlch\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742795 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.742811 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.743415 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-logs\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.743873 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2e103b-d279-4a6f-bf32-985a0389730c-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.747492 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.747774 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.750978 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.750997 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.762924 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.768957 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rv66\" (UniqueName: \"kubernetes.io/projected/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-kube-api-access-4rv66\") pod \"watcher-kuttl-api-0\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.769019 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhlch\" (UniqueName: \"kubernetes.io/projected/1b2e103b-d279-4a6f-bf32-985a0389730c-kube-api-access-jhlch\") pod \"watcher-kuttl-applier-0\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.844542 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e6b57e-42c7-4e03-8f1e-813a66d93644-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.844590 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.844623 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.844647 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4hg\" (UniqueName: \"kubernetes.io/projected/62e6b57e-42c7-4e03-8f1e-813a66d93644-kube-api-access-qd4hg\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.844696 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.845672 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e6b57e-42c7-4e03-8f1e-813a66d93644-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.848665 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.848885 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.849604 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.861083 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4hg\" (UniqueName: \"kubernetes.io/projected/62e6b57e-42c7-4e03-8f1e-813a66d93644-kube-api-access-qd4hg\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.864918 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.900321 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:39 crc kubenswrapper[4843]: I0314 09:47:39.958991 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:40 crc kubenswrapper[4843]: I0314 09:47:40.166554 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:40 crc kubenswrapper[4843]: W0314 09:47:40.177987 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b2e103b_d279_4a6f_bf32_985a0389730c.slice/crio-b8ffd794244d940854f47a551239f2a1aec4c32088c16104847e336feef934f0 WatchSource:0}: Error finding container b8ffd794244d940854f47a551239f2a1aec4c32088c16104847e336feef934f0: Status 404 returned error can't find the container with id b8ffd794244d940854f47a551239f2a1aec4c32088c16104847e336feef934f0 Mar 14 09:47:40 crc kubenswrapper[4843]: I0314 09:47:40.214788 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"1b2e103b-d279-4a6f-bf32-985a0389730c","Type":"ContainerStarted","Data":"b8ffd794244d940854f47a551239f2a1aec4c32088c16104847e336feef934f0"} Mar 14 09:47:40 crc kubenswrapper[4843]: I0314 09:47:40.515781 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:40 crc kubenswrapper[4843]: I0314 09:47:40.546521 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:40 crc kubenswrapper[4843]: W0314 09:47:40.557425 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62e6b57e_42c7_4e03_8f1e_813a66d93644.slice/crio-de94b07d5df967a5673dd16b06ce3948fcf9983d6560e09c78e60bd43365a70c WatchSource:0}: Error finding container de94b07d5df967a5673dd16b06ce3948fcf9983d6560e09c78e60bd43365a70c: Status 404 returned error can't find the container with id de94b07d5df967a5673dd16b06ce3948fcf9983d6560e09c78e60bd43365a70c Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.226621 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"62e6b57e-42c7-4e03-8f1e-813a66d93644","Type":"ContainerStarted","Data":"2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2"} Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.226906 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"62e6b57e-42c7-4e03-8f1e-813a66d93644","Type":"ContainerStarted","Data":"de94b07d5df967a5673dd16b06ce3948fcf9983d6560e09c78e60bd43365a70c"} Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.227871 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"1b2e103b-d279-4a6f-bf32-985a0389730c","Type":"ContainerStarted","Data":"97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663"} Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.229911 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85","Type":"ContainerStarted","Data":"ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6"} Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.229946 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85","Type":"ContainerStarted","Data":"461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b"} Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.229958 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85","Type":"ContainerStarted","Data":"af0ed37ed2f94f9478fbb81ce83580bc318f71eb8133434224ab6e3e4aad3e68"} Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.232750 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.251546 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.251529451 podStartE2EDuration="2.251529451s" podCreationTimestamp="2026-03-14 09:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:47:41.249808641 +0000 UTC m=+2228.562419759" watchObservedRunningTime="2026-03-14 09:47:41.251529451 +0000 UTC m=+2228.564140579" Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.283641 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.283624137 podStartE2EDuration="2.283624137s" podCreationTimestamp="2026-03-14 09:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:47:41.279928799 +0000 UTC m=+2228.592539927" watchObservedRunningTime="2026-03-14 09:47:41.283624137 +0000 UTC m=+2228.596235265" Mar 14 09:47:41 crc kubenswrapper[4843]: I0314 09:47:41.310258 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.310241951 podStartE2EDuration="2.310241951s" podCreationTimestamp="2026-03-14 09:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:47:41.305184851 +0000 UTC m=+2228.617795979" watchObservedRunningTime="2026-03-14 09:47:41.310241951 +0000 UTC m=+2228.622853079" Mar 14 09:47:43 crc kubenswrapper[4843]: I0314 09:47:43.242314 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:47:43 crc kubenswrapper[4843]: I0314 09:47:43.478058 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:44 crc kubenswrapper[4843]: I0314 09:47:44.865937 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:44 crc kubenswrapper[4843]: I0314 09:47:44.901445 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:45 crc kubenswrapper[4843]: I0314 09:47:45.387082 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:47:45 crc kubenswrapper[4843]: I0314 09:47:45.387397 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:47:45 crc kubenswrapper[4843]: I0314 09:47:45.387435 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:47:45 crc kubenswrapper[4843]: I0314 09:47:45.388197 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:47:45 crc kubenswrapper[4843]: I0314 09:47:45.388263 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" gracePeriod=600 Mar 14 09:47:45 crc kubenswrapper[4843]: E0314 09:47:45.507746 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:47:46 crc kubenswrapper[4843]: I0314 09:47:46.267234 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" exitCode=0 Mar 14 09:47:46 crc kubenswrapper[4843]: I0314 09:47:46.267289 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba"} Mar 14 09:47:46 crc kubenswrapper[4843]: I0314 09:47:46.267323 4843 scope.go:117] "RemoveContainer" containerID="1d86bf833c241f55a04a898386b9dd563f1910bdd2ccbe6124f8566a802dfff7" Mar 14 09:47:46 crc kubenswrapper[4843]: I0314 09:47:46.269953 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:47:46 crc kubenswrapper[4843]: E0314 09:47:46.270256 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:47:49 crc kubenswrapper[4843]: I0314 09:47:49.865301 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:49 crc kubenswrapper[4843]: I0314 09:47:49.893465 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:49 crc kubenswrapper[4843]: I0314 09:47:49.901704 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:49 crc kubenswrapper[4843]: I0314 09:47:49.908668 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:49 crc kubenswrapper[4843]: I0314 09:47:49.959617 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:49 crc kubenswrapper[4843]: I0314 09:47:49.990504 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:50 crc kubenswrapper[4843]: I0314 09:47:50.307523 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:50 crc kubenswrapper[4843]: I0314 09:47:50.315569 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:50 crc kubenswrapper[4843]: I0314 09:47:50.330303 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:50 crc kubenswrapper[4843]: I0314 09:47:50.350702 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:52 crc kubenswrapper[4843]: I0314 09:47:52.559720 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:52 crc kubenswrapper[4843]: I0314 09:47:52.560763 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="proxy-httpd" containerID="cri-o://6c7c6e0ba8581c41e5aba5676bf98300c6d7e98edc891045f5e1ad64d90e2106" gracePeriod=30 Mar 14 09:47:52 crc kubenswrapper[4843]: I0314 09:47:52.561285 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="sg-core" containerID="cri-o://bdb38ab86dde4597b71f9e291c0b65ef2455275a8379e5c0dfd81aa3bd49a4fe" gracePeriod=30 Mar 14 09:47:52 crc kubenswrapper[4843]: I0314 09:47:52.561356 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-notification-agent" containerID="cri-o://421d18b2d5b4387373e03df685083b5e4f778c57ff9740e308f1a0faaddb43aa" gracePeriod=30 Mar 14 09:47:52 crc kubenswrapper[4843]: I0314 09:47:52.562335 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-central-agent" containerID="cri-o://730b02f435cd77dcc4150583e231f1d1082e0efe72e629aa440264b8dacdddbf" gracePeriod=30 Mar 14 09:47:52 crc kubenswrapper[4843]: I0314 09:47:52.664334 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.160:3000/\": read tcp 10.217.0.2:45992->10.217.0.160:3000: read: connection reset by peer" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.331886 4843 generic.go:334] "Generic (PLEG): container finished" podID="0881ee30-9845-468c-8f39-581f429e437d" containerID="6c7c6e0ba8581c41e5aba5676bf98300c6d7e98edc891045f5e1ad64d90e2106" exitCode=0 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.332171 4843 generic.go:334] "Generic (PLEG): container finished" podID="0881ee30-9845-468c-8f39-581f429e437d" containerID="bdb38ab86dde4597b71f9e291c0b65ef2455275a8379e5c0dfd81aa3bd49a4fe" exitCode=2 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.332258 4843 generic.go:334] "Generic (PLEG): container finished" podID="0881ee30-9845-468c-8f39-581f429e437d" containerID="730b02f435cd77dcc4150583e231f1d1082e0efe72e629aa440264b8dacdddbf" exitCode=0 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.331916 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerDied","Data":"6c7c6e0ba8581c41e5aba5676bf98300c6d7e98edc891045f5e1ad64d90e2106"} Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.332440 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerDied","Data":"bdb38ab86dde4597b71f9e291c0b65ef2455275a8379e5c0dfd81aa3bd49a4fe"} Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.332516 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerDied","Data":"730b02f435cd77dcc4150583e231f1d1082e0efe72e629aa440264b8dacdddbf"} Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.542328 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.550059 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-lbwt5"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.607581 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.607805 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-kuttl-api-log" containerID="cri-o://461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b" gracePeriod=30 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.608247 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-api" containerID="cri-o://ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6" gracePeriod=30 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.658351 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcherb745-account-delete-ssxxz"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.659451 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.673743 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherb745-account-delete-ssxxz"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.700284 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flrdz\" (UniqueName: \"kubernetes.io/projected/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-kube-api-access-flrdz\") pod \"watcherb745-account-delete-ssxxz\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.700359 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-operator-scripts\") pod \"watcherb745-account-delete-ssxxz\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.740952 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.741209 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="62e6b57e-42c7-4e03-8f1e-813a66d93644" containerName="watcher-decision-engine" containerID="cri-o://2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2" gracePeriod=30 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.765473 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.765714 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="1b2e103b-d279-4a6f-bf32-985a0389730c" containerName="watcher-applier" containerID="cri-o://97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" gracePeriod=30 Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.802448 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flrdz\" (UniqueName: \"kubernetes.io/projected/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-kube-api-access-flrdz\") pod \"watcherb745-account-delete-ssxxz\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.802519 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-operator-scripts\") pod \"watcherb745-account-delete-ssxxz\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.803555 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-operator-scripts\") pod \"watcherb745-account-delete-ssxxz\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.834993 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flrdz\" (UniqueName: \"kubernetes.io/projected/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-kube-api-access-flrdz\") pod \"watcherb745-account-delete-ssxxz\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:53 crc kubenswrapper[4843]: I0314 09:47:53.988640 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.346873 4843 generic.go:334] "Generic (PLEG): container finished" podID="0881ee30-9845-468c-8f39-581f429e437d" containerID="421d18b2d5b4387373e03df685083b5e4f778c57ff9740e308f1a0faaddb43aa" exitCode=0 Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.347062 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerDied","Data":"421d18b2d5b4387373e03df685083b5e4f778c57ff9740e308f1a0faaddb43aa"} Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.364295 4843 generic.go:334] "Generic (PLEG): container finished" podID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerID="461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b" exitCode=143 Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.364331 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85","Type":"ContainerDied","Data":"461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b"} Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.391591 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522030 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-ceilometer-tls-certs\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522118 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qrwn\" (UniqueName: \"kubernetes.io/projected/0881ee30-9845-468c-8f39-581f429e437d-kube-api-access-4qrwn\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522207 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-log-httpd\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522246 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-config-data\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522319 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-sg-core-conf-yaml\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522830 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-scripts\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522867 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-combined-ca-bundle\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.522912 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-run-httpd\") pod \"0881ee30-9845-468c-8f39-581f429e437d\" (UID: \"0881ee30-9845-468c-8f39-581f429e437d\") " Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.523569 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.523743 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.525839 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0881ee30-9845-468c-8f39-581f429e437d-kube-api-access-4qrwn" (OuterVolumeSpecName: "kube-api-access-4qrwn") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "kube-api-access-4qrwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.526468 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-scripts" (OuterVolumeSpecName: "scripts") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.567313 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.585059 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherb745-account-delete-ssxxz"] Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.593144 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.612197 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.616908 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-config-data" (OuterVolumeSpecName: "config-data") pod "0881ee30-9845-468c-8f39-581f429e437d" (UID: "0881ee30-9845-468c-8f39-581f429e437d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.624959 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.624993 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.625002 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.625012 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.625042 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.625051 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0881ee30-9845-468c-8f39-581f429e437d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.625059 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0881ee30-9845-468c-8f39-581f429e437d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.625067 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qrwn\" (UniqueName: \"kubernetes.io/projected/0881ee30-9845-468c-8f39-581f429e437d-kube-api-access-4qrwn\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:54 crc kubenswrapper[4843]: E0314 09:47:54.867791 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:47:54 crc kubenswrapper[4843]: E0314 09:47:54.869114 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:47:54 crc kubenswrapper[4843]: E0314 09:47:54.870386 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:47:54 crc kubenswrapper[4843]: E0314 09:47:54.870458 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="1b2e103b-d279-4a6f-bf32-985a0389730c" containerName="watcher-applier" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.901651 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9322/\": dial tcp 10.217.0.165:9322: connect: connection refused" Mar 14 09:47:54 crc kubenswrapper[4843]: I0314 09:47:54.901702 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.165:9322/\": dial tcp 10.217.0.165:9322: connect: connection refused" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.210460 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.334487 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-config-data\") pod \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.334550 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-custom-prometheus-ca\") pod \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.334767 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rv66\" (UniqueName: \"kubernetes.io/projected/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-kube-api-access-4rv66\") pod \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.334800 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-combined-ca-bundle\") pod \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.334838 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-logs\") pod \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\" (UID: \"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85\") " Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.335353 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-logs" (OuterVolumeSpecName: "logs") pod "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" (UID: "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.335737 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.343482 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-kube-api-access-4rv66" (OuterVolumeSpecName: "kube-api-access-4rv66") pod "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" (UID: "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85"). InnerVolumeSpecName "kube-api-access-4rv66". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.351090 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f4ffebb-2312-4d73-8409-8cfb8ac69041" path="/var/lib/kubelet/pods/3f4ffebb-2312-4d73-8409-8cfb8ac69041/volumes" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.364173 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" (UID: "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.374433 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" (UID: "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.377254 4843 generic.go:334] "Generic (PLEG): container finished" podID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerID="ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6" exitCode=0 Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.377412 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.386924 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85","Type":"ContainerDied","Data":"ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6"} Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.386970 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85","Type":"ContainerDied","Data":"af0ed37ed2f94f9478fbb81ce83580bc318f71eb8133434224ab6e3e4aad3e68"} Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.386988 4843 scope.go:117] "RemoveContainer" containerID="ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.387952 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" containerID="58aca775f2e0644db2b9fc28254a813d0792d753edb7ace172581e32735e9b01" exitCode=0 Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.388042 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" event={"ID":"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e","Type":"ContainerDied","Data":"58aca775f2e0644db2b9fc28254a813d0792d753edb7ace172581e32735e9b01"} Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.388090 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" event={"ID":"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e","Type":"ContainerStarted","Data":"bb61dc7f8f4790aeaf67b34c896907403422e6e2ab43cbfdc6933ea90ac05b22"} Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.395919 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0881ee30-9845-468c-8f39-581f429e437d","Type":"ContainerDied","Data":"50b9debb09719ffd1a50fd05cea3b9c3d70018862d4b35a76ebe2bee463e495e"} Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.396028 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.420477 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-config-data" (OuterVolumeSpecName: "config-data") pod "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" (UID: "bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.420629 4843 scope.go:117] "RemoveContainer" containerID="461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.433521 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.439392 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.439422 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.439432 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rv66\" (UniqueName: \"kubernetes.io/projected/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-kube-api-access-4rv66\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.439440 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.453402 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.454226 4843 scope.go:117] "RemoveContainer" containerID="ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.454660 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6\": container with ID starting with ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6 not found: ID does not exist" containerID="ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.454700 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6"} err="failed to get container status \"ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6\": rpc error: code = NotFound desc = could not find container \"ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6\": container with ID starting with ba5c036cdc06f40943c886616318a84383312b8231e8f15a56463b5908de58d6 not found: ID does not exist" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.454727 4843 scope.go:117] "RemoveContainer" containerID="461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.454964 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b\": container with ID starting with 461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b not found: ID does not exist" containerID="461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.454999 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b"} err="failed to get container status \"461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b\": rpc error: code = NotFound desc = could not find container \"461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b\": container with ID starting with 461ce216fe565dca6ed6123c18d3a26cfc0fa40e54f9ad83032bf02dc1b25a0b not found: ID does not exist" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.455011 4843 scope.go:117] "RemoveContainer" containerID="6c7c6e0ba8581c41e5aba5676bf98300c6d7e98edc891045f5e1ad64d90e2106" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.473607 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.473947 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-kuttl-api-log" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.473962 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-kuttl-api-log" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.473983 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="sg-core" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.473990 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="sg-core" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.473999 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-central-agent" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474005 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-central-agent" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.474015 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-api" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474020 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-api" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.474036 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="proxy-httpd" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474041 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="proxy-httpd" Mar 14 09:47:55 crc kubenswrapper[4843]: E0314 09:47:55.474058 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-notification-agent" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474063 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-notification-agent" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474196 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="sg-core" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474205 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="proxy-httpd" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474215 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-api" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474225 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" containerName="watcher-kuttl-api-log" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474238 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-notification-agent" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.474250 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0881ee30-9845-468c-8f39-581f429e437d" containerName="ceilometer-central-agent" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.475688 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.479914 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.480148 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.480215 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.484326 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.492516 4843 scope.go:117] "RemoveContainer" containerID="bdb38ab86dde4597b71f9e291c0b65ef2455275a8379e5c0dfd81aa3bd49a4fe" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.525416 4843 scope.go:117] "RemoveContainer" containerID="421d18b2d5b4387373e03df685083b5e4f778c57ff9740e308f1a0faaddb43aa" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542473 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-config-data\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542536 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-scripts\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542561 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-run-httpd\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542615 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542653 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542692 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-log-httpd\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542724 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.542777 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkc6d\" (UniqueName: \"kubernetes.io/projected/862054d4-19a4-4fc9-b22b-f2e176c304f7-kube-api-access-zkc6d\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.551437 4843 scope.go:117] "RemoveContainer" containerID="730b02f435cd77dcc4150583e231f1d1082e0efe72e629aa440264b8dacdddbf" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644162 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644233 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkc6d\" (UniqueName: \"kubernetes.io/projected/862054d4-19a4-4fc9-b22b-f2e176c304f7-kube-api-access-zkc6d\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644298 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-config-data\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644326 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-scripts\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644345 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-run-httpd\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644384 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644411 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.644436 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-log-httpd\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.645124 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-log-httpd\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.646792 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-run-httpd\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.650395 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.650663 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-config-data\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.651485 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-scripts\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.652798 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.652867 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.663650 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkc6d\" (UniqueName: \"kubernetes.io/projected/862054d4-19a4-4fc9-b22b-f2e176c304f7-kube-api-access-zkc6d\") pod \"ceilometer-0\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.706956 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.714747 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:47:55 crc kubenswrapper[4843]: I0314 09:47:55.794182 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.314319 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.421074 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerStarted","Data":"0c9f5dbce40e37513176237ce32f93baadc3d256da441a04498a5fcf9c0b461f"} Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.694227 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.762561 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flrdz\" (UniqueName: \"kubernetes.io/projected/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-kube-api-access-flrdz\") pod \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.762689 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-operator-scripts\") pod \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\" (UID: \"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e\") " Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.763724 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" (UID: "0c7e954b-0080-4c7c-9b3d-e3b7498baf4e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.768340 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-kube-api-access-flrdz" (OuterVolumeSpecName: "kube-api-access-flrdz") pod "0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" (UID: "0c7e954b-0080-4c7c-9b3d-e3b7498baf4e"). InnerVolumeSpecName "kube-api-access-flrdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.864625 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flrdz\" (UniqueName: \"kubernetes.io/projected/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-kube-api-access-flrdz\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.864916 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:56 crc kubenswrapper[4843]: I0314 09:47:56.939512 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.071664 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhlch\" (UniqueName: \"kubernetes.io/projected/1b2e103b-d279-4a6f-bf32-985a0389730c-kube-api-access-jhlch\") pod \"1b2e103b-d279-4a6f-bf32-985a0389730c\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.071747 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-config-data\") pod \"1b2e103b-d279-4a6f-bf32-985a0389730c\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.071818 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2e103b-d279-4a6f-bf32-985a0389730c-logs\") pod \"1b2e103b-d279-4a6f-bf32-985a0389730c\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.071939 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-combined-ca-bundle\") pod \"1b2e103b-d279-4a6f-bf32-985a0389730c\" (UID: \"1b2e103b-d279-4a6f-bf32-985a0389730c\") " Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.072261 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2e103b-d279-4a6f-bf32-985a0389730c-logs" (OuterVolumeSpecName: "logs") pod "1b2e103b-d279-4a6f-bf32-985a0389730c" (UID: "1b2e103b-d279-4a6f-bf32-985a0389730c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.078547 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2e103b-d279-4a6f-bf32-985a0389730c-kube-api-access-jhlch" (OuterVolumeSpecName: "kube-api-access-jhlch") pod "1b2e103b-d279-4a6f-bf32-985a0389730c" (UID: "1b2e103b-d279-4a6f-bf32-985a0389730c"). InnerVolumeSpecName "kube-api-access-jhlch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.114605 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b2e103b-d279-4a6f-bf32-985a0389730c" (UID: "1b2e103b-d279-4a6f-bf32-985a0389730c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.121099 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.153478 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-config-data" (OuterVolumeSpecName: "config-data") pod "1b2e103b-d279-4a6f-bf32-985a0389730c" (UID: "1b2e103b-d279-4a6f-bf32-985a0389730c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.173597 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2e103b-d279-4a6f-bf32-985a0389730c-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.173644 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.173661 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhlch\" (UniqueName: \"kubernetes.io/projected/1b2e103b-d279-4a6f-bf32-985a0389730c-kube-api-access-jhlch\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.173702 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2e103b-d279-4a6f-bf32-985a0389730c-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.346889 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0881ee30-9845-468c-8f39-581f429e437d" path="/var/lib/kubelet/pods/0881ee30-9845-468c-8f39-581f429e437d/volumes" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.347879 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85" path="/var/lib/kubelet/pods/bd3bd6bc-3ac6-4b4a-9bc8-e13251cbea85/volumes" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.431978 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.432010 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherb745-account-delete-ssxxz" event={"ID":"0c7e954b-0080-4c7c-9b3d-e3b7498baf4e","Type":"ContainerDied","Data":"bb61dc7f8f4790aeaf67b34c896907403422e6e2ab43cbfdc6933ea90ac05b22"} Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.432816 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb61dc7f8f4790aeaf67b34c896907403422e6e2ab43cbfdc6933ea90ac05b22" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.435617 4843 generic.go:334] "Generic (PLEG): container finished" podID="1b2e103b-d279-4a6f-bf32-985a0389730c" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" exitCode=0 Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.435713 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"1b2e103b-d279-4a6f-bf32-985a0389730c","Type":"ContainerDied","Data":"97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663"} Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.435743 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"1b2e103b-d279-4a6f-bf32-985a0389730c","Type":"ContainerDied","Data":"b8ffd794244d940854f47a551239f2a1aec4c32088c16104847e336feef934f0"} Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.435760 4843 scope.go:117] "RemoveContainer" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.435879 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.437982 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerStarted","Data":"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31"} Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.457531 4843 scope.go:117] "RemoveContainer" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" Mar 14 09:47:57 crc kubenswrapper[4843]: E0314 09:47:57.457932 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663\": container with ID starting with 97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663 not found: ID does not exist" containerID="97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.457967 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663"} err="failed to get container status \"97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663\": rpc error: code = NotFound desc = could not find container \"97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663\": container with ID starting with 97b1b79a819488ad00b84fa31b82ccd35978bce42a17ade08c2e95607629d663 not found: ID does not exist" Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.460481 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:57 crc kubenswrapper[4843]: I0314 09:47:57.467692 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.426049 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.447524 4843 generic.go:334] "Generic (PLEG): container finished" podID="62e6b57e-42c7-4e03-8f1e-813a66d93644" containerID="2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2" exitCode=0 Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.447595 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.447620 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"62e6b57e-42c7-4e03-8f1e-813a66d93644","Type":"ContainerDied","Data":"2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2"} Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.448083 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"62e6b57e-42c7-4e03-8f1e-813a66d93644","Type":"ContainerDied","Data":"de94b07d5df967a5673dd16b06ce3948fcf9983d6560e09c78e60bd43365a70c"} Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.448128 4843 scope.go:117] "RemoveContainer" containerID="2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.455503 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerStarted","Data":"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c"} Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.455544 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerStarted","Data":"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0"} Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.469611 4843 scope.go:117] "RemoveContainer" containerID="2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2" Mar 14 09:47:58 crc kubenswrapper[4843]: E0314 09:47:58.470112 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2\": container with ID starting with 2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2 not found: ID does not exist" containerID="2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.470153 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2"} err="failed to get container status \"2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2\": rpc error: code = NotFound desc = could not find container \"2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2\": container with ID starting with 2312f3457c6a1f43940ed33a0d9f3962bc8075a93e790b39c9ab1e3455955be2 not found: ID does not exist" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.494698 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4hg\" (UniqueName: \"kubernetes.io/projected/62e6b57e-42c7-4e03-8f1e-813a66d93644-kube-api-access-qd4hg\") pod \"62e6b57e-42c7-4e03-8f1e-813a66d93644\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.494767 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-config-data\") pod \"62e6b57e-42c7-4e03-8f1e-813a66d93644\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.494864 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-custom-prometheus-ca\") pod \"62e6b57e-42c7-4e03-8f1e-813a66d93644\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.494895 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-combined-ca-bundle\") pod \"62e6b57e-42c7-4e03-8f1e-813a66d93644\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.494936 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e6b57e-42c7-4e03-8f1e-813a66d93644-logs\") pod \"62e6b57e-42c7-4e03-8f1e-813a66d93644\" (UID: \"62e6b57e-42c7-4e03-8f1e-813a66d93644\") " Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.495594 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e6b57e-42c7-4e03-8f1e-813a66d93644-logs" (OuterVolumeSpecName: "logs") pod "62e6b57e-42c7-4e03-8f1e-813a66d93644" (UID: "62e6b57e-42c7-4e03-8f1e-813a66d93644"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.505055 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e6b57e-42c7-4e03-8f1e-813a66d93644-kube-api-access-qd4hg" (OuterVolumeSpecName: "kube-api-access-qd4hg") pod "62e6b57e-42c7-4e03-8f1e-813a66d93644" (UID: "62e6b57e-42c7-4e03-8f1e-813a66d93644"). InnerVolumeSpecName "kube-api-access-qd4hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.516120 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62e6b57e-42c7-4e03-8f1e-813a66d93644" (UID: "62e6b57e-42c7-4e03-8f1e-813a66d93644"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.517799 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "62e6b57e-42c7-4e03-8f1e-813a66d93644" (UID: "62e6b57e-42c7-4e03-8f1e-813a66d93644"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.541566 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-config-data" (OuterVolumeSpecName: "config-data") pod "62e6b57e-42c7-4e03-8f1e-813a66d93644" (UID: "62e6b57e-42c7-4e03-8f1e-813a66d93644"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.596676 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4hg\" (UniqueName: \"kubernetes.io/projected/62e6b57e-42c7-4e03-8f1e-813a66d93644-kube-api-access-qd4hg\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.596722 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.596736 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.596747 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e6b57e-42c7-4e03-8f1e-813a66d93644-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.596758 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e6b57e-42c7-4e03-8f1e-813a66d93644-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.650939 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vrl6s"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.657130 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-vrl6s"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.667222 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-b745-account-create-update-l5lcz"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.673384 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcherb745-account-delete-ssxxz"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.681098 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-b745-account-create-update-l5lcz"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.687829 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcherb745-account-delete-ssxxz"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.785628 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:58 crc kubenswrapper[4843]: I0314 09:47:58.794477 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.339864 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:47:59 crc kubenswrapper[4843]: E0314 09:47:59.340463 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.349051 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" path="/var/lib/kubelet/pods/0c7e954b-0080-4c7c-9b3d-e3b7498baf4e/volumes" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.349913 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2e103b-d279-4a6f-bf32-985a0389730c" path="/var/lib/kubelet/pods/1b2e103b-d279-4a6f-bf32-985a0389730c/volumes" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.350629 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e6b57e-42c7-4e03-8f1e-813a66d93644" path="/var/lib/kubelet/pods/62e6b57e-42c7-4e03-8f1e-813a66d93644/volumes" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.351583 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f774281-e77b-4fcd-a025-1888c65e2c33" path="/var/lib/kubelet/pods/7f774281-e77b-4fcd-a025-1888c65e2c33/volumes" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.352113 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2040650-3240-48f6-a544-4041d9230929" path="/var/lib/kubelet/pods/b2040650-3240-48f6-a544-4041d9230929/volumes" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370028 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-9c25z"] Mar 14 09:47:59 crc kubenswrapper[4843]: E0314 09:47:59.370391 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e6b57e-42c7-4e03-8f1e-813a66d93644" containerName="watcher-decision-engine" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370408 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e6b57e-42c7-4e03-8f1e-813a66d93644" containerName="watcher-decision-engine" Mar 14 09:47:59 crc kubenswrapper[4843]: E0314 09:47:59.370434 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" containerName="mariadb-account-delete" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370442 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" containerName="mariadb-account-delete" Mar 14 09:47:59 crc kubenswrapper[4843]: E0314 09:47:59.370456 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2e103b-d279-4a6f-bf32-985a0389730c" containerName="watcher-applier" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370464 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2e103b-d279-4a6f-bf32-985a0389730c" containerName="watcher-applier" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370626 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e6b57e-42c7-4e03-8f1e-813a66d93644" containerName="watcher-decision-engine" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370647 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7e954b-0080-4c7c-9b3d-e3b7498baf4e" containerName="mariadb-account-delete" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.370657 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2e103b-d279-4a6f-bf32-985a0389730c" containerName="watcher-applier" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.371228 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.376229 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-9c25z"] Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.469786 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn"] Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.470780 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.474223 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.478139 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn"] Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.509255 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k8zp\" (UniqueName: \"kubernetes.io/projected/1d059309-2b46-42f5-b0b7-2a2cba0742b5-kube-api-access-8k8zp\") pod \"watcher-db-create-9c25z\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.509311 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d059309-2b46-42f5-b0b7-2a2cba0742b5-operator-scripts\") pod \"watcher-db-create-9c25z\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.610629 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfcce89b-28dd-44e6-b544-db6cc6147cc5-operator-scripts\") pod \"watcher-3f8f-account-create-update-vcqbn\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.611001 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdf82\" (UniqueName: \"kubernetes.io/projected/dfcce89b-28dd-44e6-b544-db6cc6147cc5-kube-api-access-bdf82\") pod \"watcher-3f8f-account-create-update-vcqbn\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.611128 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k8zp\" (UniqueName: \"kubernetes.io/projected/1d059309-2b46-42f5-b0b7-2a2cba0742b5-kube-api-access-8k8zp\") pod \"watcher-db-create-9c25z\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.611159 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d059309-2b46-42f5-b0b7-2a2cba0742b5-operator-scripts\") pod \"watcher-db-create-9c25z\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.611779 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d059309-2b46-42f5-b0b7-2a2cba0742b5-operator-scripts\") pod \"watcher-db-create-9c25z\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.634446 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k8zp\" (UniqueName: \"kubernetes.io/projected/1d059309-2b46-42f5-b0b7-2a2cba0742b5-kube-api-access-8k8zp\") pod \"watcher-db-create-9c25z\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.686111 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.713043 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfcce89b-28dd-44e6-b544-db6cc6147cc5-operator-scripts\") pod \"watcher-3f8f-account-create-update-vcqbn\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.713125 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdf82\" (UniqueName: \"kubernetes.io/projected/dfcce89b-28dd-44e6-b544-db6cc6147cc5-kube-api-access-bdf82\") pod \"watcher-3f8f-account-create-update-vcqbn\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.713963 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfcce89b-28dd-44e6-b544-db6cc6147cc5-operator-scripts\") pod \"watcher-3f8f-account-create-update-vcqbn\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.734291 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdf82\" (UniqueName: \"kubernetes.io/projected/dfcce89b-28dd-44e6-b544-db6cc6147cc5-kube-api-access-bdf82\") pod \"watcher-3f8f-account-create-update-vcqbn\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:47:59 crc kubenswrapper[4843]: I0314 09:47:59.820137 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.487972 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558028-b8mw2"] Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.491971 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.496149 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.496357 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.496555 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.516313 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558028-b8mw2"] Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.525442 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qlbl\" (UniqueName: \"kubernetes.io/projected/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b-kube-api-access-6qlbl\") pod \"auto-csr-approver-29558028-b8mw2\" (UID: \"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b\") " pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.535359 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-9c25z"] Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.551442 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn"] Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.626769 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qlbl\" (UniqueName: \"kubernetes.io/projected/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b-kube-api-access-6qlbl\") pod \"auto-csr-approver-29558028-b8mw2\" (UID: \"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b\") " pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.644935 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qlbl\" (UniqueName: \"kubernetes.io/projected/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b-kube-api-access-6qlbl\") pod \"auto-csr-approver-29558028-b8mw2\" (UID: \"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b\") " pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:00 crc kubenswrapper[4843]: I0314 09:48:00.836279 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:01 crc kubenswrapper[4843]: W0314 09:48:01.491912 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f0f2ca8_7ffe_4d0f_8138_b6e25849a29b.slice/crio-be61a2fdad8354b64d9326ce0fc04d0cfea968de6c87b5c5f66b43ee025148bf WatchSource:0}: Error finding container be61a2fdad8354b64d9326ce0fc04d0cfea968de6c87b5c5f66b43ee025148bf: Status 404 returned error can't find the container with id be61a2fdad8354b64d9326ce0fc04d0cfea968de6c87b5c5f66b43ee025148bf Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.493999 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerStarted","Data":"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf"} Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.494219 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-central-agent" containerID="cri-o://d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" gracePeriod=30 Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.494461 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.494564 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="proxy-httpd" containerID="cri-o://1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" gracePeriod=30 Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.494819 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="sg-core" containerID="cri-o://eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" gracePeriod=30 Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.494889 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-notification-agent" containerID="cri-o://509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" gracePeriod=30 Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.502728 4843 generic.go:334] "Generic (PLEG): container finished" podID="dfcce89b-28dd-44e6-b544-db6cc6147cc5" containerID="086156d7f2f3de1323812bef986b946e34740a54f026f70c9aec24f3506df087" exitCode=0 Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.502794 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" event={"ID":"dfcce89b-28dd-44e6-b544-db6cc6147cc5","Type":"ContainerDied","Data":"086156d7f2f3de1323812bef986b946e34740a54f026f70c9aec24f3506df087"} Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.502824 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" event={"ID":"dfcce89b-28dd-44e6-b544-db6cc6147cc5","Type":"ContainerStarted","Data":"3d7c92aaab51bb0132316290ea184a94ee28ef8d8fb224393161c06674969fc0"} Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.508820 4843 generic.go:334] "Generic (PLEG): container finished" podID="1d059309-2b46-42f5-b0b7-2a2cba0742b5" containerID="6efbdf372a841f2ec509a824c2acd7bb46cdacdc022eab68f0bb7ec7dc9c8b23" exitCode=0 Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.508873 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-9c25z" event={"ID":"1d059309-2b46-42f5-b0b7-2a2cba0742b5","Type":"ContainerDied","Data":"6efbdf372a841f2ec509a824c2acd7bb46cdacdc022eab68f0bb7ec7dc9c8b23"} Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.508905 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-9c25z" event={"ID":"1d059309-2b46-42f5-b0b7-2a2cba0742b5","Type":"ContainerStarted","Data":"17c47de0b6540919e544c7ee40e1c724514504d7d9fb4816f29a699bc29df10a"} Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.520089 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558028-b8mw2"] Mar 14 09:48:01 crc kubenswrapper[4843]: I0314 09:48:01.526499 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=3.01222275 podStartE2EDuration="6.526479679s" podCreationTimestamp="2026-03-14 09:47:55 +0000 UTC" firstStartedPulling="2026-03-14 09:47:56.331403334 +0000 UTC m=+2243.644014462" lastFinishedPulling="2026-03-14 09:47:59.845660263 +0000 UTC m=+2247.158271391" observedRunningTime="2026-03-14 09:48:01.514986065 +0000 UTC m=+2248.827597203" watchObservedRunningTime="2026-03-14 09:48:01.526479679 +0000 UTC m=+2248.839090817" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.283397 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355589 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkc6d\" (UniqueName: \"kubernetes.io/projected/862054d4-19a4-4fc9-b22b-f2e176c304f7-kube-api-access-zkc6d\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355717 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-run-httpd\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355753 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-ceilometer-tls-certs\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355791 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-log-httpd\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355830 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-scripts\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355889 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-sg-core-conf-yaml\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.355945 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-config-data\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.356004 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-combined-ca-bundle\") pod \"862054d4-19a4-4fc9-b22b-f2e176c304f7\" (UID: \"862054d4-19a4-4fc9-b22b-f2e176c304f7\") " Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.356096 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.356766 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.357034 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.357059 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/862054d4-19a4-4fc9-b22b-f2e176c304f7-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.361147 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862054d4-19a4-4fc9-b22b-f2e176c304f7-kube-api-access-zkc6d" (OuterVolumeSpecName: "kube-api-access-zkc6d") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "kube-api-access-zkc6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.361462 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-scripts" (OuterVolumeSpecName: "scripts") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.383450 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.407898 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.417093 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.446460 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-config-data" (OuterVolumeSpecName: "config-data") pod "862054d4-19a4-4fc9-b22b-f2e176c304f7" (UID: "862054d4-19a4-4fc9-b22b-f2e176c304f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.458996 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.459031 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.459045 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkc6d\" (UniqueName: \"kubernetes.io/projected/862054d4-19a4-4fc9-b22b-f2e176c304f7-kube-api-access-zkc6d\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.459058 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.459067 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.459077 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/862054d4-19a4-4fc9-b22b-f2e176c304f7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.518715 4843 generic.go:334] "Generic (PLEG): container finished" podID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" exitCode=0 Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.519915 4843 generic.go:334] "Generic (PLEG): container finished" podID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" exitCode=2 Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.519997 4843 generic.go:334] "Generic (PLEG): container finished" podID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" exitCode=0 Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.520057 4843 generic.go:334] "Generic (PLEG): container finished" podID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" exitCode=0 Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.518789 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.518773 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerDied","Data":"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf"} Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.520335 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerDied","Data":"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c"} Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.520362 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerDied","Data":"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0"} Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.520377 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerDied","Data":"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31"} Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.520389 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"862054d4-19a4-4fc9-b22b-f2e176c304f7","Type":"ContainerDied","Data":"0c9f5dbce40e37513176237ce32f93baadc3d256da441a04498a5fcf9c0b461f"} Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.520409 4843 scope.go:117] "RemoveContainer" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.522192 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" event={"ID":"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b","Type":"ContainerStarted","Data":"be61a2fdad8354b64d9326ce0fc04d0cfea968de6c87b5c5f66b43ee025148bf"} Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.557692 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.562149 4843 scope.go:117] "RemoveContainer" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.571581 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.582551 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.582860 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="sg-core" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.582877 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="sg-core" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.582897 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-central-agent" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.582904 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-central-agent" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.582919 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-notification-agent" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.582925 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-notification-agent" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.582935 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="proxy-httpd" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.582941 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="proxy-httpd" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.583438 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-notification-agent" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.583461 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="sg-core" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.583471 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="ceilometer-central-agent" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.583481 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" containerName="proxy-httpd" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.584740 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.592120 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.592415 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.592585 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.601872 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.610673 4843 scope.go:117] "RemoveContainer" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.657637 4843 scope.go:117] "RemoveContainer" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664706 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-scripts\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664771 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-log-httpd\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664824 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-config-data\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664890 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664914 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw22x\" (UniqueName: \"kubernetes.io/projected/a2fad06f-b2de-4f84-a613-3275945e91d5-kube-api-access-sw22x\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664936 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664952 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-run-httpd\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.664969 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.683757 4843 scope.go:117] "RemoveContainer" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.684723 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": container with ID starting with 1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf not found: ID does not exist" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.684777 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf"} err="failed to get container status \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": rpc error: code = NotFound desc = could not find container \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": container with ID starting with 1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.684813 4843 scope.go:117] "RemoveContainer" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.686415 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": container with ID starting with eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c not found: ID does not exist" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.686456 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c"} err="failed to get container status \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": rpc error: code = NotFound desc = could not find container \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": container with ID starting with eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.686479 4843 scope.go:117] "RemoveContainer" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.686907 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": container with ID starting with 509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0 not found: ID does not exist" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.686935 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0"} err="failed to get container status \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": rpc error: code = NotFound desc = could not find container \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": container with ID starting with 509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.686952 4843 scope.go:117] "RemoveContainer" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" Mar 14 09:48:02 crc kubenswrapper[4843]: E0314 09:48:02.687219 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": container with ID starting with d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31 not found: ID does not exist" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.687249 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31"} err="failed to get container status \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": rpc error: code = NotFound desc = could not find container \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": container with ID starting with d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.687262 4843 scope.go:117] "RemoveContainer" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.687573 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf"} err="failed to get container status \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": rpc error: code = NotFound desc = could not find container \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": container with ID starting with 1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.687594 4843 scope.go:117] "RemoveContainer" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.687813 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c"} err="failed to get container status \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": rpc error: code = NotFound desc = could not find container \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": container with ID starting with eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.687832 4843 scope.go:117] "RemoveContainer" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.688033 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0"} err="failed to get container status \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": rpc error: code = NotFound desc = could not find container \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": container with ID starting with 509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.688044 4843 scope.go:117] "RemoveContainer" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.688247 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31"} err="failed to get container status \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": rpc error: code = NotFound desc = could not find container \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": container with ID starting with d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.688258 4843 scope.go:117] "RemoveContainer" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.692383 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf"} err="failed to get container status \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": rpc error: code = NotFound desc = could not find container \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": container with ID starting with 1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.692425 4843 scope.go:117] "RemoveContainer" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.692948 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c"} err="failed to get container status \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": rpc error: code = NotFound desc = could not find container \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": container with ID starting with eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.692983 4843 scope.go:117] "RemoveContainer" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.693206 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0"} err="failed to get container status \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": rpc error: code = NotFound desc = could not find container \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": container with ID starting with 509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.693229 4843 scope.go:117] "RemoveContainer" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.693462 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31"} err="failed to get container status \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": rpc error: code = NotFound desc = could not find container \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": container with ID starting with d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.693484 4843 scope.go:117] "RemoveContainer" containerID="1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.695751 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf"} err="failed to get container status \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": rpc error: code = NotFound desc = could not find container \"1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf\": container with ID starting with 1ba58157d6e7a8d98980b8d5972de98220914b9af71107e632824e0911b6eccf not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.695789 4843 scope.go:117] "RemoveContainer" containerID="eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.696117 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c"} err="failed to get container status \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": rpc error: code = NotFound desc = could not find container \"eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c\": container with ID starting with eb72a700121b4db6f7123fcf671a1983c4e0617b44661ed0da4401820c454a7c not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.696143 4843 scope.go:117] "RemoveContainer" containerID="509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.697332 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0"} err="failed to get container status \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": rpc error: code = NotFound desc = could not find container \"509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0\": container with ID starting with 509786cd726b41cdec90b82bff909629470f2fb96cff1c9872438e63dff6dde0 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.697358 4843 scope.go:117] "RemoveContainer" containerID="d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.698024 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31"} err="failed to get container status \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": rpc error: code = NotFound desc = could not find container \"d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31\": container with ID starting with d165545d0548a36214bb74de4362e7b2200292c27d5a0ccbf443304035341d31 not found: ID does not exist" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.766856 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.767659 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw22x\" (UniqueName: \"kubernetes.io/projected/a2fad06f-b2de-4f84-a613-3275945e91d5-kube-api-access-sw22x\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.767702 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.767721 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-run-httpd\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.767746 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.767816 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-scripts\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.767842 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-log-httpd\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.768169 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-log-httpd\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.768653 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-config-data\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.770435 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-run-httpd\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.771714 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.774295 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.774836 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.777327 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-scripts\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.778179 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-config-data\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.795339 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw22x\" (UniqueName: \"kubernetes.io/projected/a2fad06f-b2de-4f84-a613-3275945e91d5-kube-api-access-sw22x\") pod \"ceilometer-0\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:02 crc kubenswrapper[4843]: I0314 09:48:02.928992 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.004615 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.022576 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.071987 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k8zp\" (UniqueName: \"kubernetes.io/projected/1d059309-2b46-42f5-b0b7-2a2cba0742b5-kube-api-access-8k8zp\") pod \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.072142 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdf82\" (UniqueName: \"kubernetes.io/projected/dfcce89b-28dd-44e6-b544-db6cc6147cc5-kube-api-access-bdf82\") pod \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.072193 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d059309-2b46-42f5-b0b7-2a2cba0742b5-operator-scripts\") pod \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\" (UID: \"1d059309-2b46-42f5-b0b7-2a2cba0742b5\") " Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.072243 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfcce89b-28dd-44e6-b544-db6cc6147cc5-operator-scripts\") pod \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\" (UID: \"dfcce89b-28dd-44e6-b544-db6cc6147cc5\") " Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.073597 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfcce89b-28dd-44e6-b544-db6cc6147cc5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfcce89b-28dd-44e6-b544-db6cc6147cc5" (UID: "dfcce89b-28dd-44e6-b544-db6cc6147cc5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.073855 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d059309-2b46-42f5-b0b7-2a2cba0742b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d059309-2b46-42f5-b0b7-2a2cba0742b5" (UID: "1d059309-2b46-42f5-b0b7-2a2cba0742b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.078630 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d059309-2b46-42f5-b0b7-2a2cba0742b5-kube-api-access-8k8zp" (OuterVolumeSpecName: "kube-api-access-8k8zp") pod "1d059309-2b46-42f5-b0b7-2a2cba0742b5" (UID: "1d059309-2b46-42f5-b0b7-2a2cba0742b5"). InnerVolumeSpecName "kube-api-access-8k8zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.080460 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfcce89b-28dd-44e6-b544-db6cc6147cc5-kube-api-access-bdf82" (OuterVolumeSpecName: "kube-api-access-bdf82") pod "dfcce89b-28dd-44e6-b544-db6cc6147cc5" (UID: "dfcce89b-28dd-44e6-b544-db6cc6147cc5"). InnerVolumeSpecName "kube-api-access-bdf82". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.174083 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdf82\" (UniqueName: \"kubernetes.io/projected/dfcce89b-28dd-44e6-b544-db6cc6147cc5-kube-api-access-bdf82\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.174460 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d059309-2b46-42f5-b0b7-2a2cba0742b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.174475 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfcce89b-28dd-44e6-b544-db6cc6147cc5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.174486 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k8zp\" (UniqueName: \"kubernetes.io/projected/1d059309-2b46-42f5-b0b7-2a2cba0742b5-kube-api-access-8k8zp\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.352316 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862054d4-19a4-4fc9-b22b-f2e176c304f7" path="/var/lib/kubelet/pods/862054d4-19a4-4fc9-b22b-f2e176c304f7/volumes" Mar 14 09:48:03 crc kubenswrapper[4843]: W0314 09:48:03.415584 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2fad06f_b2de_4f84_a613_3275945e91d5.slice/crio-e193b1cbbde76c90563c09fb9c6b4d3d12efdeaf078846cc29c0181b07d4642e WatchSource:0}: Error finding container e193b1cbbde76c90563c09fb9c6b4d3d12efdeaf078846cc29c0181b07d4642e: Status 404 returned error can't find the container with id e193b1cbbde76c90563c09fb9c6b4d3d12efdeaf078846cc29c0181b07d4642e Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.419672 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.533938 4843 generic.go:334] "Generic (PLEG): container finished" podID="7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b" containerID="4559a397bebfabe03c183b58325b36823c289dd3bad2c87f701bf5905c15bb2d" exitCode=0 Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.534004 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" event={"ID":"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b","Type":"ContainerDied","Data":"4559a397bebfabe03c183b58325b36823c289dd3bad2c87f701bf5905c15bb2d"} Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.536513 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-9c25z" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.537218 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-9c25z" event={"ID":"1d059309-2b46-42f5-b0b7-2a2cba0742b5","Type":"ContainerDied","Data":"17c47de0b6540919e544c7ee40e1c724514504d7d9fb4816f29a699bc29df10a"} Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.537250 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17c47de0b6540919e544c7ee40e1c724514504d7d9fb4816f29a699bc29df10a" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.544835 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerStarted","Data":"e193b1cbbde76c90563c09fb9c6b4d3d12efdeaf078846cc29c0181b07d4642e"} Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.553162 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.553072 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn" event={"ID":"dfcce89b-28dd-44e6-b544-db6cc6147cc5","Type":"ContainerDied","Data":"3d7c92aaab51bb0132316290ea184a94ee28ef8d8fb224393161c06674969fc0"} Mar 14 09:48:03 crc kubenswrapper[4843]: I0314 09:48:03.553917 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d7c92aaab51bb0132316290ea184a94ee28ef8d8fb224393161c06674969fc0" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.562658 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerStarted","Data":"bd97b22d9f69670932bd06d957fbb5dda0be7f83ee5221a13d3a08c0632a877e"} Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.828245 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j"] Mar 14 09:48:04 crc kubenswrapper[4843]: E0314 09:48:04.828567 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d059309-2b46-42f5-b0b7-2a2cba0742b5" containerName="mariadb-database-create" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.828582 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d059309-2b46-42f5-b0b7-2a2cba0742b5" containerName="mariadb-database-create" Mar 14 09:48:04 crc kubenswrapper[4843]: E0314 09:48:04.828610 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfcce89b-28dd-44e6-b544-db6cc6147cc5" containerName="mariadb-account-create-update" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.828616 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfcce89b-28dd-44e6-b544-db6cc6147cc5" containerName="mariadb-account-create-update" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.828753 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfcce89b-28dd-44e6-b544-db6cc6147cc5" containerName="mariadb-account-create-update" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.828772 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d059309-2b46-42f5-b0b7-2a2cba0742b5" containerName="mariadb-database-create" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.829244 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.834155 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-cbqff" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.834417 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.842553 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j"] Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.902114 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-config-data\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.902223 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.902255 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qf8l\" (UniqueName: \"kubernetes.io/projected/489c8399-7b8e-47e1-9483-90fcd8a81ba2-kube-api-access-9qf8l\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.902309 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:04 crc kubenswrapper[4843]: I0314 09:48:04.967132 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.004493 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qlbl\" (UniqueName: \"kubernetes.io/projected/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b-kube-api-access-6qlbl\") pod \"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b\" (UID: \"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b\") " Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.004754 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-config-data\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.004799 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.004834 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qf8l\" (UniqueName: \"kubernetes.io/projected/489c8399-7b8e-47e1-9483-90fcd8a81ba2-kube-api-access-9qf8l\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.004865 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.010655 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.010679 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-config-data\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.015972 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.017945 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b-kube-api-access-6qlbl" (OuterVolumeSpecName: "kube-api-access-6qlbl") pod "7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b" (UID: "7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b"). InnerVolumeSpecName "kube-api-access-6qlbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.028735 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qf8l\" (UniqueName: \"kubernetes.io/projected/489c8399-7b8e-47e1-9483-90fcd8a81ba2-kube-api-access-9qf8l\") pod \"watcher-kuttl-db-sync-rrv7j\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.106416 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qlbl\" (UniqueName: \"kubernetes.io/projected/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b-kube-api-access-6qlbl\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.151065 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.571063 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" event={"ID":"7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b","Type":"ContainerDied","Data":"be61a2fdad8354b64d9326ce0fc04d0cfea968de6c87b5c5f66b43ee025148bf"} Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.571089 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558028-b8mw2" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.571101 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be61a2fdad8354b64d9326ce0fc04d0cfea968de6c87b5c5f66b43ee025148bf" Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.573144 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerStarted","Data":"22186a6a166b31b87ae5ee3238b93308d80a71a0e84ed9f30ffb441f8cf846db"} Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.573166 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerStarted","Data":"d0fb0eacf2d2e2ea95c820e567d501dbd5ebdf5f3fc794657dd69a0b38966fa5"} Mar 14 09:48:05 crc kubenswrapper[4843]: I0314 09:48:05.710336 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j"] Mar 14 09:48:06 crc kubenswrapper[4843]: I0314 09:48:06.018830 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558022-g74p7"] Mar 14 09:48:06 crc kubenswrapper[4843]: I0314 09:48:06.024659 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558022-g74p7"] Mar 14 09:48:06 crc kubenswrapper[4843]: I0314 09:48:06.583213 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" event={"ID":"489c8399-7b8e-47e1-9483-90fcd8a81ba2","Type":"ContainerStarted","Data":"da4b54856f14ad3b8dd74f9c2d512a81dbf3786bd07508c800043ded8f64892f"} Mar 14 09:48:06 crc kubenswrapper[4843]: I0314 09:48:06.583283 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" event={"ID":"489c8399-7b8e-47e1-9483-90fcd8a81ba2","Type":"ContainerStarted","Data":"86770c7ab89ce115bd828e014edbc7fa30af56bd4858e1a99340bdebc596ac64"} Mar 14 09:48:07 crc kubenswrapper[4843]: I0314 09:48:07.348177 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ed054ee-d91b-43df-ba81-b0abbe047bc1" path="/var/lib/kubelet/pods/3ed054ee-d91b-43df-ba81-b0abbe047bc1/volumes" Mar 14 09:48:07 crc kubenswrapper[4843]: I0314 09:48:07.593080 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerStarted","Data":"981d5136fbb20e997b37b6f65833623696c2abf8040d351eab890e0cd4ce9e67"} Mar 14 09:48:07 crc kubenswrapper[4843]: I0314 09:48:07.625360 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.093844548 podStartE2EDuration="5.625324018s" podCreationTimestamp="2026-03-14 09:48:02 +0000 UTC" firstStartedPulling="2026-03-14 09:48:03.418221892 +0000 UTC m=+2250.730833020" lastFinishedPulling="2026-03-14 09:48:06.949701362 +0000 UTC m=+2254.262312490" observedRunningTime="2026-03-14 09:48:07.616489138 +0000 UTC m=+2254.929100276" watchObservedRunningTime="2026-03-14 09:48:07.625324018 +0000 UTC m=+2254.937935196" Mar 14 09:48:07 crc kubenswrapper[4843]: I0314 09:48:07.630353 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" podStartSLOduration=3.6303354580000002 podStartE2EDuration="3.630335458s" podCreationTimestamp="2026-03-14 09:48:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:06.599434793 +0000 UTC m=+2253.912045931" watchObservedRunningTime="2026-03-14 09:48:07.630335458 +0000 UTC m=+2254.942946626" Mar 14 09:48:08 crc kubenswrapper[4843]: I0314 09:48:08.602371 4843 generic.go:334] "Generic (PLEG): container finished" podID="489c8399-7b8e-47e1-9483-90fcd8a81ba2" containerID="da4b54856f14ad3b8dd74f9c2d512a81dbf3786bd07508c800043ded8f64892f" exitCode=0 Mar 14 09:48:08 crc kubenswrapper[4843]: I0314 09:48:08.602419 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" event={"ID":"489c8399-7b8e-47e1-9483-90fcd8a81ba2","Type":"ContainerDied","Data":"da4b54856f14ad3b8dd74f9c2d512a81dbf3786bd07508c800043ded8f64892f"} Mar 14 09:48:08 crc kubenswrapper[4843]: I0314 09:48:08.602855 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:09 crc kubenswrapper[4843]: I0314 09:48:09.948325 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.078366 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-db-sync-config-data\") pod \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.078481 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-config-data\") pod \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.078627 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qf8l\" (UniqueName: \"kubernetes.io/projected/489c8399-7b8e-47e1-9483-90fcd8a81ba2-kube-api-access-9qf8l\") pod \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.078693 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-combined-ca-bundle\") pod \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\" (UID: \"489c8399-7b8e-47e1-9483-90fcd8a81ba2\") " Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.084496 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489c8399-7b8e-47e1-9483-90fcd8a81ba2-kube-api-access-9qf8l" (OuterVolumeSpecName: "kube-api-access-9qf8l") pod "489c8399-7b8e-47e1-9483-90fcd8a81ba2" (UID: "489c8399-7b8e-47e1-9483-90fcd8a81ba2"). InnerVolumeSpecName "kube-api-access-9qf8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.098457 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "489c8399-7b8e-47e1-9483-90fcd8a81ba2" (UID: "489c8399-7b8e-47e1-9483-90fcd8a81ba2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.102828 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "489c8399-7b8e-47e1-9483-90fcd8a81ba2" (UID: "489c8399-7b8e-47e1-9483-90fcd8a81ba2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.156441 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-config-data" (OuterVolumeSpecName: "config-data") pod "489c8399-7b8e-47e1-9483-90fcd8a81ba2" (UID: "489c8399-7b8e-47e1-9483-90fcd8a81ba2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.180187 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qf8l\" (UniqueName: \"kubernetes.io/projected/489c8399-7b8e-47e1-9483-90fcd8a81ba2-kube-api-access-9qf8l\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.180219 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.180228 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.180239 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/489c8399-7b8e-47e1-9483-90fcd8a81ba2-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.651010 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" event={"ID":"489c8399-7b8e-47e1-9483-90fcd8a81ba2","Type":"ContainerDied","Data":"86770c7ab89ce115bd828e014edbc7fa30af56bd4858e1a99340bdebc596ac64"} Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.651068 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86770c7ab89ce115bd828e014edbc7fa30af56bd4858e1a99340bdebc596ac64" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.651163 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.943239 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:10 crc kubenswrapper[4843]: E0314 09:48:10.943673 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b" containerName="oc" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.943695 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b" containerName="oc" Mar 14 09:48:10 crc kubenswrapper[4843]: E0314 09:48:10.943734 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489c8399-7b8e-47e1-9483-90fcd8a81ba2" containerName="watcher-kuttl-db-sync" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.943743 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="489c8399-7b8e-47e1-9483-90fcd8a81ba2" containerName="watcher-kuttl-db-sync" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.943926 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b" containerName="oc" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.943959 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="489c8399-7b8e-47e1-9483-90fcd8a81ba2" containerName="watcher-kuttl-db-sync" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.944719 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.946731 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-cbqff" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.946963 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.959469 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.975738 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.977028 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.982661 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.991871 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31a93d6e-3696-4d5a-aa25-5fab7407743e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.994441 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.994686 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq92r\" (UniqueName: \"kubernetes.io/projected/310a9747-2adf-41dc-8210-874357c5e5ac-kube-api-access-tq92r\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.994805 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310a9747-2adf-41dc-8210-874357c5e5ac-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.994932 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.995032 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbkb\" (UniqueName: \"kubernetes.io/projected/31a93d6e-3696-4d5a-aa25-5fab7407743e-kube-api-access-trbkb\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.995131 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.995306 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:10 crc kubenswrapper[4843]: I0314 09:48:10.995459 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.021329 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.063946 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.065560 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.070989 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097011 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq92r\" (UniqueName: \"kubernetes.io/projected/310a9747-2adf-41dc-8210-874357c5e5ac-kube-api-access-tq92r\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097056 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310a9747-2adf-41dc-8210-874357c5e5ac-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097072 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097088 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trbkb\" (UniqueName: \"kubernetes.io/projected/31a93d6e-3696-4d5a-aa25-5fab7407743e-kube-api-access-trbkb\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097105 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097128 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-logs\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097162 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097182 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097199 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097225 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097254 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31a93d6e-3696-4d5a-aa25-5fab7407743e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097288 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6hf\" (UniqueName: \"kubernetes.io/projected/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-kube-api-access-5n6hf\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097315 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.097332 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.098406 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310a9747-2adf-41dc-8210-874357c5e5ac-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.098994 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31a93d6e-3696-4d5a-aa25-5fab7407743e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.101524 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.101541 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.101735 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.102108 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.108776 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.113614 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbkb\" (UniqueName: \"kubernetes.io/projected/31a93d6e-3696-4d5a-aa25-5fab7407743e-kube-api-access-trbkb\") pod \"watcher-kuttl-applier-0\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.116261 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq92r\" (UniqueName: \"kubernetes.io/projected/310a9747-2adf-41dc-8210-874357c5e5ac-kube-api-access-tq92r\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.132429 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.197945 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-logs\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.198023 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.198077 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.198106 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n6hf\" (UniqueName: \"kubernetes.io/projected/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-kube-api-access-5n6hf\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.198128 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.200707 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-logs\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.202813 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.205164 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.206746 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.217796 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n6hf\" (UniqueName: \"kubernetes.io/projected/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-kube-api-access-5n6hf\") pod \"watcher-kuttl-api-0\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.277877 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.305505 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.481784 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.846913 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:11 crc kubenswrapper[4843]: I0314 09:48:11.878565 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:11 crc kubenswrapper[4843]: W0314 09:48:11.883935 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31a93d6e_3696_4d5a_aa25_5fab7407743e.slice/crio-ab7632168d0477b377e206a3f7169df79b5417318fb6024ab043124374745bf3 WatchSource:0}: Error finding container ab7632168d0477b377e206a3f7169df79b5417318fb6024ab043124374745bf3: Status 404 returned error can't find the container with id ab7632168d0477b377e206a3f7169df79b5417318fb6024ab043124374745bf3 Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.001128 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.667465 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"31a93d6e-3696-4d5a-aa25-5fab7407743e","Type":"ContainerStarted","Data":"27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.667504 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"31a93d6e-3696-4d5a-aa25-5fab7407743e","Type":"ContainerStarted","Data":"ab7632168d0477b377e206a3f7169df79b5417318fb6024ab043124374745bf3"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.670444 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"3dc9e11a-5f77-4ebb-b179-8999c7b6c038","Type":"ContainerStarted","Data":"41e1a3cc450175bdd3c66ab8dbab87623b4b036b22120573179a28d33fb38143"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.670520 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.670535 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"3dc9e11a-5f77-4ebb-b179-8999c7b6c038","Type":"ContainerStarted","Data":"168c03c90765333edfe189348e8244d95f01fe682f669a2220ac042257465ca9"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.670549 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"3dc9e11a-5f77-4ebb-b179-8999c7b6c038","Type":"ContainerStarted","Data":"5d8114d0edf6641954aea8889fd51a623810e3946d1af9714a503b7841a48700"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.672197 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"310a9747-2adf-41dc-8210-874357c5e5ac","Type":"ContainerStarted","Data":"ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.672240 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"310a9747-2adf-41dc-8210-874357c5e5ac","Type":"ContainerStarted","Data":"98f87202208e8df23f3e2aabced1c5686f55b5f332362e28fb86da7163ac0383"} Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.689109 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.6890907840000002 podStartE2EDuration="2.689090784s" podCreationTimestamp="2026-03-14 09:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:12.685894477 +0000 UTC m=+2259.998505605" watchObservedRunningTime="2026-03-14 09:48:12.689090784 +0000 UTC m=+2260.001701912" Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.731340 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.731324661 podStartE2EDuration="2.731324661s" podCreationTimestamp="2026-03-14 09:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:12.730234464 +0000 UTC m=+2260.042845592" watchObservedRunningTime="2026-03-14 09:48:12.731324661 +0000 UTC m=+2260.043935789" Mar 14 09:48:12 crc kubenswrapper[4843]: I0314 09:48:12.736053 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=1.7360356829999999 podStartE2EDuration="1.736035683s" podCreationTimestamp="2026-03-14 09:48:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:12.712062431 +0000 UTC m=+2260.024673559" watchObservedRunningTime="2026-03-14 09:48:12.736035683 +0000 UTC m=+2260.048646811" Mar 14 09:48:13 crc kubenswrapper[4843]: I0314 09:48:13.344093 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:48:13 crc kubenswrapper[4843]: E0314 09:48:13.344385 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:48:14 crc kubenswrapper[4843]: I0314 09:48:14.992704 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:16 crc kubenswrapper[4843]: I0314 09:48:16.306362 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:16 crc kubenswrapper[4843]: I0314 09:48:16.482915 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.279014 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.304704 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.306182 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.355962 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.483552 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.488027 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.756970 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.761393 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.808186 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:21 crc kubenswrapper[4843]: I0314 09:48:21.837528 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:23 crc kubenswrapper[4843]: I0314 09:48:23.910488 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:23 crc kubenswrapper[4843]: I0314 09:48:23.911403 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-central-agent" containerID="cri-o://bd97b22d9f69670932bd06d957fbb5dda0be7f83ee5221a13d3a08c0632a877e" gracePeriod=30 Mar 14 09:48:23 crc kubenswrapper[4843]: I0314 09:48:23.911428 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="proxy-httpd" containerID="cri-o://981d5136fbb20e997b37b6f65833623696c2abf8040d351eab890e0cd4ce9e67" gracePeriod=30 Mar 14 09:48:23 crc kubenswrapper[4843]: I0314 09:48:23.911467 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="sg-core" containerID="cri-o://22186a6a166b31b87ae5ee3238b93308d80a71a0e84ed9f30ffb441f8cf846db" gracePeriod=30 Mar 14 09:48:23 crc kubenswrapper[4843]: I0314 09:48:23.911477 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-notification-agent" containerID="cri-o://d0fb0eacf2d2e2ea95c820e567d501dbd5ebdf5f3fc794657dd69a0b38966fa5" gracePeriod=30 Mar 14 09:48:23 crc kubenswrapper[4843]: I0314 09:48:23.926341 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.172:3000/\": EOF" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.140681 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.153094 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-rrv7j"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.199460 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher3f8f-account-delete-m8b86"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.200510 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.212488 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher3f8f-account-delete-m8b86"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.233164 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.304639 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trrc4\" (UniqueName: \"kubernetes.io/projected/0242bec3-8402-414b-af6d-1c58175dcc70-kube-api-access-trrc4\") pod \"watcher3f8f-account-delete-m8b86\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.304768 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0242bec3-8402-414b-af6d-1c58175dcc70-operator-scripts\") pod \"watcher3f8f-account-delete-m8b86\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.322905 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.323106 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-kuttl-api-log" containerID="cri-o://168c03c90765333edfe189348e8244d95f01fe682f669a2220ac042257465ca9" gracePeriod=30 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.323251 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-api" containerID="cri-o://41e1a3cc450175bdd3c66ab8dbab87623b4b036b22120573179a28d33fb38143" gracePeriod=30 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.339069 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:48:24 crc kubenswrapper[4843]: E0314 09:48:24.339425 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.350264 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.350473 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="31a93d6e-3696-4d5a-aa25-5fab7407743e" containerName="watcher-applier" containerID="cri-o://27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" gracePeriod=30 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.405753 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0242bec3-8402-414b-af6d-1c58175dcc70-operator-scripts\") pod \"watcher3f8f-account-delete-m8b86\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.405895 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trrc4\" (UniqueName: \"kubernetes.io/projected/0242bec3-8402-414b-af6d-1c58175dcc70-kube-api-access-trrc4\") pod \"watcher3f8f-account-delete-m8b86\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.407681 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0242bec3-8402-414b-af6d-1c58175dcc70-operator-scripts\") pod \"watcher3f8f-account-delete-m8b86\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.430504 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trrc4\" (UniqueName: \"kubernetes.io/projected/0242bec3-8402-414b-af6d-1c58175dcc70-kube-api-access-trrc4\") pod \"watcher3f8f-account-delete-m8b86\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.519861 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.812063 4843 generic.go:334] "Generic (PLEG): container finished" podID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerID="168c03c90765333edfe189348e8244d95f01fe682f669a2220ac042257465ca9" exitCode=143 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.812374 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"3dc9e11a-5f77-4ebb-b179-8999c7b6c038","Type":"ContainerDied","Data":"168c03c90765333edfe189348e8244d95f01fe682f669a2220ac042257465ca9"} Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.832708 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerID="981d5136fbb20e997b37b6f65833623696c2abf8040d351eab890e0cd4ce9e67" exitCode=0 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.832740 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerID="22186a6a166b31b87ae5ee3238b93308d80a71a0e84ed9f30ffb441f8cf846db" exitCode=2 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.832748 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerID="d0fb0eacf2d2e2ea95c820e567d501dbd5ebdf5f3fc794657dd69a0b38966fa5" exitCode=0 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.832755 4843 generic.go:334] "Generic (PLEG): container finished" podID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerID="bd97b22d9f69670932bd06d957fbb5dda0be7f83ee5221a13d3a08c0632a877e" exitCode=0 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.832927 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="310a9747-2adf-41dc-8210-874357c5e5ac" containerName="watcher-decision-engine" containerID="cri-o://ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519" gracePeriod=30 Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.833203 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerDied","Data":"981d5136fbb20e997b37b6f65833623696c2abf8040d351eab890e0cd4ce9e67"} Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.833284 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerDied","Data":"22186a6a166b31b87ae5ee3238b93308d80a71a0e84ed9f30ffb441f8cf846db"} Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.833299 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerDied","Data":"d0fb0eacf2d2e2ea95c820e567d501dbd5ebdf5f3fc794657dd69a0b38966fa5"} Mar 14 09:48:24 crc kubenswrapper[4843]: I0314 09:48:24.833311 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerDied","Data":"bd97b22d9f69670932bd06d957fbb5dda0be7f83ee5221a13d3a08c0632a877e"} Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.170177 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:25 crc kubenswrapper[4843]: W0314 09:48:25.261304 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0242bec3_8402_414b_af6d_1c58175dcc70.slice/crio-e92e8c8b7c4e97cb9067558073ee5fd590d85bde78fd1996a2134a439bd6bf5c WatchSource:0}: Error finding container e92e8c8b7c4e97cb9067558073ee5fd590d85bde78fd1996a2134a439bd6bf5c: Status 404 returned error can't find the container with id e92e8c8b7c4e97cb9067558073ee5fd590d85bde78fd1996a2134a439bd6bf5c Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.275810 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher3f8f-account-delete-m8b86"] Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341171 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-run-httpd\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341229 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-combined-ca-bundle\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341356 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-log-httpd\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341402 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw22x\" (UniqueName: \"kubernetes.io/projected/a2fad06f-b2de-4f84-a613-3275945e91d5-kube-api-access-sw22x\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341427 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-scripts\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341472 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-ceilometer-tls-certs\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341493 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-config-data\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.341591 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-sg-core-conf-yaml\") pod \"a2fad06f-b2de-4f84-a613-3275945e91d5\" (UID: \"a2fad06f-b2de-4f84-a613-3275945e91d5\") " Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.342576 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.345088 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.348521 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2fad06f-b2de-4f84-a613-3275945e91d5-kube-api-access-sw22x" (OuterVolumeSpecName: "kube-api-access-sw22x") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "kube-api-access-sw22x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.353708 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="489c8399-7b8e-47e1-9483-90fcd8a81ba2" path="/var/lib/kubelet/pods/489c8399-7b8e-47e1-9483-90fcd8a81ba2/volumes" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.355634 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-scripts" (OuterVolumeSpecName: "scripts") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.375204 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.414316 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.439093 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.445844 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.446149 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.446162 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.446174 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a2fad06f-b2de-4f84-a613-3275945e91d5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.446185 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw22x\" (UniqueName: \"kubernetes.io/projected/a2fad06f-b2de-4f84-a613-3275945e91d5-kube-api-access-sw22x\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.446197 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.446210 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.464731 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-config-data" (OuterVolumeSpecName: "config-data") pod "a2fad06f-b2de-4f84-a613-3275945e91d5" (UID: "a2fad06f-b2de-4f84-a613-3275945e91d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.547919 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fad06f-b2de-4f84-a613-3275945e91d5-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.873289 4843 generic.go:334] "Generic (PLEG): container finished" podID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerID="41e1a3cc450175bdd3c66ab8dbab87623b4b036b22120573179a28d33fb38143" exitCode=0 Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.873351 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"3dc9e11a-5f77-4ebb-b179-8999c7b6c038","Type":"ContainerDied","Data":"41e1a3cc450175bdd3c66ab8dbab87623b4b036b22120573179a28d33fb38143"} Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.878650 4843 generic.go:334] "Generic (PLEG): container finished" podID="0242bec3-8402-414b-af6d-1c58175dcc70" containerID="81815b8418cdf4f5b7d4d8f9dc00bc6db437d8a38581b33df5df65168f303900" exitCode=0 Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.878739 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" event={"ID":"0242bec3-8402-414b-af6d-1c58175dcc70","Type":"ContainerDied","Data":"81815b8418cdf4f5b7d4d8f9dc00bc6db437d8a38581b33df5df65168f303900"} Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.878765 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" event={"ID":"0242bec3-8402-414b-af6d-1c58175dcc70","Type":"ContainerStarted","Data":"e92e8c8b7c4e97cb9067558073ee5fd590d85bde78fd1996a2134a439bd6bf5c"} Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.893630 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.893786 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a2fad06f-b2de-4f84-a613-3275945e91d5","Type":"ContainerDied","Data":"e193b1cbbde76c90563c09fb9c6b4d3d12efdeaf078846cc29c0181b07d4642e"} Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.893828 4843 scope.go:117] "RemoveContainer" containerID="981d5136fbb20e997b37b6f65833623696c2abf8040d351eab890e0cd4ce9e67" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.893947 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.938137 4843 scope.go:117] "RemoveContainer" containerID="22186a6a166b31b87ae5ee3238b93308d80a71a0e84ed9f30ffb441f8cf846db" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.971018 4843 scope.go:117] "RemoveContainer" containerID="d0fb0eacf2d2e2ea95c820e567d501dbd5ebdf5f3fc794657dd69a0b38966fa5" Mar 14 09:48:25 crc kubenswrapper[4843]: I0314 09:48:25.989393 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.011972 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.023598 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.024003 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="proxy-httpd" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024019 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="proxy-httpd" Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.024028 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-api" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024034 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-api" Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.024064 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-notification-agent" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024071 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-notification-agent" Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.024086 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="sg-core" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024092 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="sg-core" Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.024100 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-kuttl-api-log" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024106 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-kuttl-api-log" Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.024113 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-central-agent" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024119 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-central-agent" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024299 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-notification-agent" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024318 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="ceilometer-central-agent" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024327 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-kuttl-api-log" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024337 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="proxy-httpd" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024348 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" containerName="sg-core" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.024358 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" containerName="watcher-api" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.025698 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.025775 4843 scope.go:117] "RemoveContainer" containerID="bd97b22d9f69670932bd06d957fbb5dda0be7f83ee5221a13d3a08c0632a877e" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.028730 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.029014 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.029127 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.034292 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.056826 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-combined-ca-bundle\") pod \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.056886 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-logs\") pod \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.056973 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-config-data\") pod \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.057082 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n6hf\" (UniqueName: \"kubernetes.io/projected/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-kube-api-access-5n6hf\") pod \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.057183 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-custom-prometheus-ca\") pod \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\" (UID: \"3dc9e11a-5f77-4ebb-b179-8999c7b6c038\") " Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.057691 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-logs" (OuterVolumeSpecName: "logs") pod "3dc9e11a-5f77-4ebb-b179-8999c7b6c038" (UID: "3dc9e11a-5f77-4ebb-b179-8999c7b6c038"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.084294 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-kube-api-access-5n6hf" (OuterVolumeSpecName: "kube-api-access-5n6hf") pod "3dc9e11a-5f77-4ebb-b179-8999c7b6c038" (UID: "3dc9e11a-5f77-4ebb-b179-8999c7b6c038"). InnerVolumeSpecName "kube-api-access-5n6hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.144517 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dc9e11a-5f77-4ebb-b179-8999c7b6c038" (UID: "3dc9e11a-5f77-4ebb-b179-8999c7b6c038"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.155816 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "3dc9e11a-5f77-4ebb-b179-8999c7b6c038" (UID: "3dc9e11a-5f77-4ebb-b179-8999c7b6c038"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159508 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-scripts\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159577 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-run-httpd\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159621 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159670 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-config-data\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159709 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159732 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpsfx\" (UniqueName: \"kubernetes.io/projected/26baba94-695b-467b-9ab8-d6c8de1824ec-kube-api-access-lpsfx\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159805 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-log-httpd\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159843 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159912 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159931 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159944 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.159956 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n6hf\" (UniqueName: \"kubernetes.io/projected/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-kube-api-access-5n6hf\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.166022 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-config-data" (OuterVolumeSpecName: "config-data") pod "3dc9e11a-5f77-4ebb-b179-8999c7b6c038" (UID: "3dc9e11a-5f77-4ebb-b179-8999c7b6c038"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261004 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-run-httpd\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261066 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261107 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-config-data\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261139 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261158 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpsfx\" (UniqueName: \"kubernetes.io/projected/26baba94-695b-467b-9ab8-d6c8de1824ec-kube-api-access-lpsfx\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261206 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-log-httpd\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261232 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261254 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-scripts\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261323 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc9e11a-5f77-4ebb-b179-8999c7b6c038-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261556 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-run-httpd\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.261839 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-log-httpd\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.264642 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-scripts\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.264705 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.267899 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.271308 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.271603 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-config-data\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.280766 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpsfx\" (UniqueName: \"kubernetes.io/projected/26baba94-695b-467b-9ab8-d6c8de1824ec-kube-api-access-lpsfx\") pod \"ceilometer-0\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.309186 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.310539 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.313909 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:48:26 crc kubenswrapper[4843]: E0314 09:48:26.313978 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="31a93d6e-3696-4d5a-aa25-5fab7407743e" containerName="watcher-applier" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.429645 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.827665 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.899123 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.907322 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.907297 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"3dc9e11a-5f77-4ebb-b179-8999c7b6c038","Type":"ContainerDied","Data":"5d8114d0edf6641954aea8889fd51a623810e3946d1af9714a503b7841a48700"} Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.907611 4843 scope.go:117] "RemoveContainer" containerID="41e1a3cc450175bdd3c66ab8dbab87623b4b036b22120573179a28d33fb38143" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.909302 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerStarted","Data":"6abb8d6c673db9db2cc96b513483a598d83d39773dedd001d60daea611dd21cf"} Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.930518 4843 scope.go:117] "RemoveContainer" containerID="168c03c90765333edfe189348e8244d95f01fe682f669a2220ac042257465ca9" Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.945630 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:26 crc kubenswrapper[4843]: I0314 09:48:26.951069 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.252965 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.349898 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dc9e11a-5f77-4ebb-b179-8999c7b6c038" path="/var/lib/kubelet/pods/3dc9e11a-5f77-4ebb-b179-8999c7b6c038/volumes" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.350628 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2fad06f-b2de-4f84-a613-3275945e91d5" path="/var/lib/kubelet/pods/a2fad06f-b2de-4f84-a613-3275945e91d5/volumes" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.385189 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trrc4\" (UniqueName: \"kubernetes.io/projected/0242bec3-8402-414b-af6d-1c58175dcc70-kube-api-access-trrc4\") pod \"0242bec3-8402-414b-af6d-1c58175dcc70\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.385352 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0242bec3-8402-414b-af6d-1c58175dcc70-operator-scripts\") pod \"0242bec3-8402-414b-af6d-1c58175dcc70\" (UID: \"0242bec3-8402-414b-af6d-1c58175dcc70\") " Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.386182 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0242bec3-8402-414b-af6d-1c58175dcc70-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0242bec3-8402-414b-af6d-1c58175dcc70" (UID: "0242bec3-8402-414b-af6d-1c58175dcc70"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.388792 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0242bec3-8402-414b-af6d-1c58175dcc70-kube-api-access-trrc4" (OuterVolumeSpecName: "kube-api-access-trrc4") pod "0242bec3-8402-414b-af6d-1c58175dcc70" (UID: "0242bec3-8402-414b-af6d-1c58175dcc70"). InnerVolumeSpecName "kube-api-access-trrc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.487197 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trrc4\" (UniqueName: \"kubernetes.io/projected/0242bec3-8402-414b-af6d-1c58175dcc70-kube-api-access-trrc4\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.487521 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0242bec3-8402-414b-af6d-1c58175dcc70-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.921383 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerStarted","Data":"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9"} Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.922597 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" event={"ID":"0242bec3-8402-414b-af6d-1c58175dcc70","Type":"ContainerDied","Data":"e92e8c8b7c4e97cb9067558073ee5fd590d85bde78fd1996a2134a439bd6bf5c"} Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.922630 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e92e8c8b7c4e97cb9067558073ee5fd590d85bde78fd1996a2134a439bd6bf5c" Mar 14 09:48:27 crc kubenswrapper[4843]: I0314 09:48:27.922680 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher3f8f-account-delete-m8b86" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.393930 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.504357 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31a93d6e-3696-4d5a-aa25-5fab7407743e-logs\") pod \"31a93d6e-3696-4d5a-aa25-5fab7407743e\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.504540 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trbkb\" (UniqueName: \"kubernetes.io/projected/31a93d6e-3696-4d5a-aa25-5fab7407743e-kube-api-access-trbkb\") pod \"31a93d6e-3696-4d5a-aa25-5fab7407743e\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.504655 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-config-data\") pod \"31a93d6e-3696-4d5a-aa25-5fab7407743e\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.504797 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31a93d6e-3696-4d5a-aa25-5fab7407743e-logs" (OuterVolumeSpecName: "logs") pod "31a93d6e-3696-4d5a-aa25-5fab7407743e" (UID: "31a93d6e-3696-4d5a-aa25-5fab7407743e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.505386 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-combined-ca-bundle\") pod \"31a93d6e-3696-4d5a-aa25-5fab7407743e\" (UID: \"31a93d6e-3696-4d5a-aa25-5fab7407743e\") " Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.506177 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31a93d6e-3696-4d5a-aa25-5fab7407743e-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.509637 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a93d6e-3696-4d5a-aa25-5fab7407743e-kube-api-access-trbkb" (OuterVolumeSpecName: "kube-api-access-trbkb") pod "31a93d6e-3696-4d5a-aa25-5fab7407743e" (UID: "31a93d6e-3696-4d5a-aa25-5fab7407743e"). InnerVolumeSpecName "kube-api-access-trbkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.528034 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31a93d6e-3696-4d5a-aa25-5fab7407743e" (UID: "31a93d6e-3696-4d5a-aa25-5fab7407743e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.550677 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-config-data" (OuterVolumeSpecName: "config-data") pod "31a93d6e-3696-4d5a-aa25-5fab7407743e" (UID: "31a93d6e-3696-4d5a-aa25-5fab7407743e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.607178 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trbkb\" (UniqueName: \"kubernetes.io/projected/31a93d6e-3696-4d5a-aa25-5fab7407743e-kube-api-access-trbkb\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.607216 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.607229 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a93d6e-3696-4d5a-aa25-5fab7407743e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.932684 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerStarted","Data":"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e"} Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.933391 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerStarted","Data":"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667"} Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.934550 4843 generic.go:334] "Generic (PLEG): container finished" podID="31a93d6e-3696-4d5a-aa25-5fab7407743e" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" exitCode=0 Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.934663 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"31a93d6e-3696-4d5a-aa25-5fab7407743e","Type":"ContainerDied","Data":"27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d"} Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.934765 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"31a93d6e-3696-4d5a-aa25-5fab7407743e","Type":"ContainerDied","Data":"ab7632168d0477b377e206a3f7169df79b5417318fb6024ab043124374745bf3"} Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.934869 4843 scope.go:117] "RemoveContainer" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.935103 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.985172 4843 scope.go:117] "RemoveContainer" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" Mar 14 09:48:28 crc kubenswrapper[4843]: E0314 09:48:28.985674 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d\": container with ID starting with 27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d not found: ID does not exist" containerID="27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.985702 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d"} err="failed to get container status \"27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d\": rpc error: code = NotFound desc = could not find container \"27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d\": container with ID starting with 27b402af37b723ea677cb3962f97fd6b5e429652a530db259f53bca1a1ee662d not found: ID does not exist" Mar 14 09:48:28 crc kubenswrapper[4843]: I0314 09:48:28.991396 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.001436 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.219825 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-9c25z"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.227171 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-9c25z"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.233560 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher3f8f-account-delete-m8b86"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.240418 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher3f8f-account-delete-m8b86"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.246530 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.252477 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-3f8f-account-create-update-vcqbn"] Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.348580 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0242bec3-8402-414b-af6d-1c58175dcc70" path="/var/lib/kubelet/pods/0242bec3-8402-414b-af6d-1c58175dcc70/volumes" Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.349323 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d059309-2b46-42f5-b0b7-2a2cba0742b5" path="/var/lib/kubelet/pods/1d059309-2b46-42f5-b0b7-2a2cba0742b5/volumes" Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.350452 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31a93d6e-3696-4d5a-aa25-5fab7407743e" path="/var/lib/kubelet/pods/31a93d6e-3696-4d5a-aa25-5fab7407743e/volumes" Mar 14 09:48:29 crc kubenswrapper[4843]: I0314 09:48:29.351262 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfcce89b-28dd-44e6-b544-db6cc6147cc5" path="/var/lib/kubelet/pods/dfcce89b-28dd-44e6-b544-db6cc6147cc5/volumes" Mar 14 09:48:30 crc kubenswrapper[4843]: I0314 09:48:30.951113 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerStarted","Data":"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b"} Mar 14 09:48:30 crc kubenswrapper[4843]: I0314 09:48:30.951406 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:30 crc kubenswrapper[4843]: I0314 09:48:30.951234 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-central-agent" containerID="cri-o://54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" gracePeriod=30 Mar 14 09:48:30 crc kubenswrapper[4843]: I0314 09:48:30.951347 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-notification-agent" containerID="cri-o://3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" gracePeriod=30 Mar 14 09:48:30 crc kubenswrapper[4843]: I0314 09:48:30.951328 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="sg-core" containerID="cri-o://02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" gracePeriod=30 Mar 14 09:48:30 crc kubenswrapper[4843]: I0314 09:48:30.951327 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="proxy-httpd" containerID="cri-o://1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" gracePeriod=30 Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.004820 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.783089402 podStartE2EDuration="6.004800708s" podCreationTimestamp="2026-03-14 09:48:25 +0000 UTC" firstStartedPulling="2026-03-14 09:48:26.896245053 +0000 UTC m=+2274.208856181" lastFinishedPulling="2026-03-14 09:48:30.117956359 +0000 UTC m=+2277.430567487" observedRunningTime="2026-03-14 09:48:31.000707931 +0000 UTC m=+2278.313319059" watchObservedRunningTime="2026-03-14 09:48:31.004800708 +0000 UTC m=+2278.317411836" Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.921355 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974428 4843 generic.go:334] "Generic (PLEG): container finished" podID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" exitCode=0 Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974461 4843 generic.go:334] "Generic (PLEG): container finished" podID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" exitCode=2 Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974468 4843 generic.go:334] "Generic (PLEG): container finished" podID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" exitCode=0 Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974476 4843 generic.go:334] "Generic (PLEG): container finished" podID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" exitCode=0 Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974494 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerDied","Data":"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b"} Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974523 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerDied","Data":"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e"} Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974538 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerDied","Data":"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667"} Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974551 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerDied","Data":"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9"} Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974562 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"26baba94-695b-467b-9ab8-d6c8de1824ec","Type":"ContainerDied","Data":"6abb8d6c673db9db2cc96b513483a598d83d39773dedd001d60daea611dd21cf"} Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974578 4843 scope.go:117] "RemoveContainer" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.974728 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:31 crc kubenswrapper[4843]: I0314 09:48:31.999179 4843 scope.go:117] "RemoveContainer" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.016244 4843 scope.go:117] "RemoveContainer" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.032096 4843 scope.go:117] "RemoveContainer" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.047827 4843 scope.go:117] "RemoveContainer" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.060053 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": container with ID starting with 1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b not found: ID does not exist" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.060103 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b"} err="failed to get container status \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": rpc error: code = NotFound desc = could not find container \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": container with ID starting with 1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.060131 4843 scope.go:117] "RemoveContainer" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.060894 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": container with ID starting with 02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e not found: ID does not exist" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.060974 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e"} err="failed to get container status \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": rpc error: code = NotFound desc = could not find container \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": container with ID starting with 02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.061008 4843 scope.go:117] "RemoveContainer" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.061542 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": container with ID starting with 3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667 not found: ID does not exist" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.061568 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667"} err="failed to get container status \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": rpc error: code = NotFound desc = could not find container \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": container with ID starting with 3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.061584 4843 scope.go:117] "RemoveContainer" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.061920 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": container with ID starting with 54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9 not found: ID does not exist" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.061955 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9"} err="failed to get container status \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": rpc error: code = NotFound desc = could not find container \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": container with ID starting with 54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.061975 4843 scope.go:117] "RemoveContainer" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.066512 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b"} err="failed to get container status \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": rpc error: code = NotFound desc = could not find container \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": container with ID starting with 1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.066555 4843 scope.go:117] "RemoveContainer" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.066923 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e"} err="failed to get container status \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": rpc error: code = NotFound desc = could not find container \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": container with ID starting with 02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.066941 4843 scope.go:117] "RemoveContainer" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067183 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667"} err="failed to get container status \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": rpc error: code = NotFound desc = could not find container \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": container with ID starting with 3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067200 4843 scope.go:117] "RemoveContainer" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067460 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9"} err="failed to get container status \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": rpc error: code = NotFound desc = could not find container \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": container with ID starting with 54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067476 4843 scope.go:117] "RemoveContainer" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067703 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b"} err="failed to get container status \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": rpc error: code = NotFound desc = could not find container \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": container with ID starting with 1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067720 4843 scope.go:117] "RemoveContainer" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067919 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e"} err="failed to get container status \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": rpc error: code = NotFound desc = could not find container \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": container with ID starting with 02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.067934 4843 scope.go:117] "RemoveContainer" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.068156 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667"} err="failed to get container status \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": rpc error: code = NotFound desc = could not find container \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": container with ID starting with 3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.068174 4843 scope.go:117] "RemoveContainer" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.068454 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9"} err="failed to get container status \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": rpc error: code = NotFound desc = could not find container \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": container with ID starting with 54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.068484 4843 scope.go:117] "RemoveContainer" containerID="1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.068752 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b"} err="failed to get container status \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": rpc error: code = NotFound desc = could not find container \"1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b\": container with ID starting with 1d250d99ae7d30fc05f1838097e69b2815e5bdf6c74493c37f58356db94a2a6b not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.068766 4843 scope.go:117] "RemoveContainer" containerID="02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.069132 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e"} err="failed to get container status \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": rpc error: code = NotFound desc = could not find container \"02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e\": container with ID starting with 02fd89ef0052a8145dbb00b541ad9bb83472a9c4baed7059527482bf3e45304e not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.069158 4843 scope.go:117] "RemoveContainer" containerID="3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.069550 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667"} err="failed to get container status \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": rpc error: code = NotFound desc = could not find container \"3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667\": container with ID starting with 3b1fd064f09f7ad8e85a6124fd1cd36bfe474a7e34fc3cdc62075a838b920667 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.069571 4843 scope.go:117] "RemoveContainer" containerID="54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.069928 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9"} err="failed to get container status \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": rpc error: code = NotFound desc = could not find container \"54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9\": container with ID starting with 54d1d57675b49795a80e7e4b5f6c3b0de1cb600a9391bc1689082bd0142090e9 not found: ID does not exist" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.074908 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpsfx\" (UniqueName: \"kubernetes.io/projected/26baba94-695b-467b-9ab8-d6c8de1824ec-kube-api-access-lpsfx\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.074973 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-ceilometer-tls-certs\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.075002 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-scripts\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.075033 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-run-httpd\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.075095 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-log-httpd\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.075133 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-sg-core-conf-yaml\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.075172 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-combined-ca-bundle\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.075236 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-config-data\") pod \"26baba94-695b-467b-9ab8-d6c8de1824ec\" (UID: \"26baba94-695b-467b-9ab8-d6c8de1824ec\") " Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.076642 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.077026 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.089481 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-scripts" (OuterVolumeSpecName: "scripts") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.089623 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26baba94-695b-467b-9ab8-d6c8de1824ec-kube-api-access-lpsfx" (OuterVolumeSpecName: "kube-api-access-lpsfx") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "kube-api-access-lpsfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.108839 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.126144 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.149842 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.163209 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-config-data" (OuterVolumeSpecName: "config-data") pod "26baba94-695b-467b-9ab8-d6c8de1824ec" (UID: "26baba94-695b-467b-9ab8-d6c8de1824ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177260 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpsfx\" (UniqueName: \"kubernetes.io/projected/26baba94-695b-467b-9ab8-d6c8de1824ec-kube-api-access-lpsfx\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177304 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177314 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177323 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177331 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26baba94-695b-467b-9ab8-d6c8de1824ec-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177339 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177348 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.177355 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26baba94-695b-467b-9ab8-d6c8de1824ec-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.314975 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.325737 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.337973 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.338357 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-central-agent" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338374 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-central-agent" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.338403 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="proxy-httpd" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338411 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="proxy-httpd" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.338428 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="sg-core" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338434 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="sg-core" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.338445 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0242bec3-8402-414b-af6d-1c58175dcc70" containerName="mariadb-account-delete" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338452 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0242bec3-8402-414b-af6d-1c58175dcc70" containerName="mariadb-account-delete" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.338462 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-notification-agent" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338469 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-notification-agent" Mar 14 09:48:32 crc kubenswrapper[4843]: E0314 09:48:32.338483 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a93d6e-3696-4d5a-aa25-5fab7407743e" containerName="watcher-applier" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338490 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a93d6e-3696-4d5a-aa25-5fab7407743e" containerName="watcher-applier" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338667 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0242bec3-8402-414b-af6d-1c58175dcc70" containerName="mariadb-account-delete" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338683 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-notification-agent" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338696 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="proxy-httpd" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338708 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a93d6e-3696-4d5a-aa25-5fab7407743e" containerName="watcher-applier" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338719 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="ceilometer-central-agent" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.338731 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" containerName="sg-core" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.340374 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.342888 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.343065 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.343188 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.361130 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.503740 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504201 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-scripts\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504234 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-log-httpd\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504258 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504354 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tczhm\" (UniqueName: \"kubernetes.io/projected/24ab212e-5406-458a-97ef-a18305940668-kube-api-access-tczhm\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504402 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-config-data\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504479 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-run-httpd\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.504611 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.605792 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-scripts\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.605838 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-log-httpd\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.605858 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.605885 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tczhm\" (UniqueName: \"kubernetes.io/projected/24ab212e-5406-458a-97ef-a18305940668-kube-api-access-tczhm\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.605910 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-config-data\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.605949 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-run-httpd\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.606002 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.606039 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.606355 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-log-httpd\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.606633 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-run-httpd\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.609968 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.610230 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.610994 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-scripts\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.612077 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.625673 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-config-data\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.631923 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tczhm\" (UniqueName: \"kubernetes.io/projected/24ab212e-5406-458a-97ef-a18305940668-kube-api-access-tczhm\") pod \"ceilometer-0\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:32 crc kubenswrapper[4843]: I0314 09:48:32.703885 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:33 crc kubenswrapper[4843]: I0314 09:48:33.152818 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:48:33 crc kubenswrapper[4843]: I0314 09:48:33.365662 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26baba94-695b-467b-9ab8-d6c8de1824ec" path="/var/lib/kubelet/pods/26baba94-695b-467b-9ab8-d6c8de1824ec/volumes" Mar 14 09:48:33 crc kubenswrapper[4843]: I0314 09:48:33.989811 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerStarted","Data":"168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f"} Mar 14 09:48:33 crc kubenswrapper[4843]: I0314 09:48:33.989861 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerStarted","Data":"f04f6fb57bb1224a8793cc6f305ceef1eeaa8a5a895f4011c748ed29747af412"} Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.816579 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.942541 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-config-data\") pod \"310a9747-2adf-41dc-8210-874357c5e5ac\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.942779 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq92r\" (UniqueName: \"kubernetes.io/projected/310a9747-2adf-41dc-8210-874357c5e5ac-kube-api-access-tq92r\") pod \"310a9747-2adf-41dc-8210-874357c5e5ac\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.942940 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310a9747-2adf-41dc-8210-874357c5e5ac-logs\") pod \"310a9747-2adf-41dc-8210-874357c5e5ac\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.943224 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/310a9747-2adf-41dc-8210-874357c5e5ac-logs" (OuterVolumeSpecName: "logs") pod "310a9747-2adf-41dc-8210-874357c5e5ac" (UID: "310a9747-2adf-41dc-8210-874357c5e5ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.943442 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-combined-ca-bundle\") pod \"310a9747-2adf-41dc-8210-874357c5e5ac\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.943777 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-custom-prometheus-ca\") pod \"310a9747-2adf-41dc-8210-874357c5e5ac\" (UID: \"310a9747-2adf-41dc-8210-874357c5e5ac\") " Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.944321 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/310a9747-2adf-41dc-8210-874357c5e5ac-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.946885 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/310a9747-2adf-41dc-8210-874357c5e5ac-kube-api-access-tq92r" (OuterVolumeSpecName: "kube-api-access-tq92r") pod "310a9747-2adf-41dc-8210-874357c5e5ac" (UID: "310a9747-2adf-41dc-8210-874357c5e5ac"). InnerVolumeSpecName "kube-api-access-tq92r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.963053 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "310a9747-2adf-41dc-8210-874357c5e5ac" (UID: "310a9747-2adf-41dc-8210-874357c5e5ac"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.969388 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "310a9747-2adf-41dc-8210-874357c5e5ac" (UID: "310a9747-2adf-41dc-8210-874357c5e5ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.981234 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-config-data" (OuterVolumeSpecName: "config-data") pod "310a9747-2adf-41dc-8210-874357c5e5ac" (UID: "310a9747-2adf-41dc-8210-874357c5e5ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.998330 4843 generic.go:334] "Generic (PLEG): container finished" podID="310a9747-2adf-41dc-8210-874357c5e5ac" containerID="ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519" exitCode=0 Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.998395 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"310a9747-2adf-41dc-8210-874357c5e5ac","Type":"ContainerDied","Data":"ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519"} Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.998420 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"310a9747-2adf-41dc-8210-874357c5e5ac","Type":"ContainerDied","Data":"98f87202208e8df23f3e2aabced1c5686f55b5f332362e28fb86da7163ac0383"} Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.998436 4843 scope.go:117] "RemoveContainer" containerID="ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519" Mar 14 09:48:34 crc kubenswrapper[4843]: I0314 09:48:34.998531 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.004106 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerStarted","Data":"0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334"} Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.022698 4843 scope.go:117] "RemoveContainer" containerID="ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519" Mar 14 09:48:35 crc kubenswrapper[4843]: E0314 09:48:35.023232 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519\": container with ID starting with ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519 not found: ID does not exist" containerID="ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.023304 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519"} err="failed to get container status \"ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519\": rpc error: code = NotFound desc = could not find container \"ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519\": container with ID starting with ebfa390f2e3f829af8825dfa519862a67e5abb759ecd19a9c669a7f524900519 not found: ID does not exist" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.034542 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.039961 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.046447 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.046472 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq92r\" (UniqueName: \"kubernetes.io/projected/310a9747-2adf-41dc-8210-874357c5e5ac-kube-api-access-tq92r\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.046482 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.046490 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/310a9747-2adf-41dc-8210-874357c5e5ac-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:35 crc kubenswrapper[4843]: I0314 09:48:35.347108 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="310a9747-2adf-41dc-8210-874357c5e5ac" path="/var/lib/kubelet/pods/310a9747-2adf-41dc-8210-874357c5e5ac/volumes" Mar 14 09:48:36 crc kubenswrapper[4843]: I0314 09:48:36.012428 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerStarted","Data":"f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9"} Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.022298 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerStarted","Data":"1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f"} Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.022681 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.056091 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.558197155 podStartE2EDuration="5.056068164s" podCreationTimestamp="2026-03-14 09:48:32 +0000 UTC" firstStartedPulling="2026-03-14 09:48:33.155770941 +0000 UTC m=+2280.468382069" lastFinishedPulling="2026-03-14 09:48:36.65364191 +0000 UTC m=+2283.966253078" observedRunningTime="2026-03-14 09:48:37.04712271 +0000 UTC m=+2284.359733838" watchObservedRunningTime="2026-03-14 09:48:37.056068164 +0000 UTC m=+2284.368679292" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.339041 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:48:37 crc kubenswrapper[4843]: E0314 09:48:37.339570 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.657059 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv"] Mar 14 09:48:37 crc kubenswrapper[4843]: E0314 09:48:37.657562 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="310a9747-2adf-41dc-8210-874357c5e5ac" containerName="watcher-decision-engine" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.657586 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="310a9747-2adf-41dc-8210-874357c5e5ac" containerName="watcher-decision-engine" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.657811 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="310a9747-2adf-41dc-8210-874357c5e5ac" containerName="watcher-decision-engine" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.658549 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.660722 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.669297 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-t9qwm"] Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.670648 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.675198 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv"] Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.681697 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-t9qwm"] Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.795188 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-operator-scripts\") pod \"watcher-5aaf-account-create-update-v5dbv\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.795846 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtrrt\" (UniqueName: \"kubernetes.io/projected/be5ed09f-5725-43cc-8ae5-fc8deac795ea-kube-api-access-xtrrt\") pod \"watcher-db-create-t9qwm\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.795966 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx9zr\" (UniqueName: \"kubernetes.io/projected/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-kube-api-access-jx9zr\") pod \"watcher-5aaf-account-create-update-v5dbv\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.796192 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be5ed09f-5725-43cc-8ae5-fc8deac795ea-operator-scripts\") pod \"watcher-db-create-t9qwm\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.898026 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtrrt\" (UniqueName: \"kubernetes.io/projected/be5ed09f-5725-43cc-8ae5-fc8deac795ea-kube-api-access-xtrrt\") pod \"watcher-db-create-t9qwm\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.898089 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx9zr\" (UniqueName: \"kubernetes.io/projected/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-kube-api-access-jx9zr\") pod \"watcher-5aaf-account-create-update-v5dbv\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.898148 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be5ed09f-5725-43cc-8ae5-fc8deac795ea-operator-scripts\") pod \"watcher-db-create-t9qwm\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.898198 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-operator-scripts\") pod \"watcher-5aaf-account-create-update-v5dbv\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.899005 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-operator-scripts\") pod \"watcher-5aaf-account-create-update-v5dbv\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.899133 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be5ed09f-5725-43cc-8ae5-fc8deac795ea-operator-scripts\") pod \"watcher-db-create-t9qwm\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.920666 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx9zr\" (UniqueName: \"kubernetes.io/projected/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-kube-api-access-jx9zr\") pod \"watcher-5aaf-account-create-update-v5dbv\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.930009 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtrrt\" (UniqueName: \"kubernetes.io/projected/be5ed09f-5725-43cc-8ae5-fc8deac795ea-kube-api-access-xtrrt\") pod \"watcher-db-create-t9qwm\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:37 crc kubenswrapper[4843]: I0314 09:48:37.975264 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:38 crc kubenswrapper[4843]: I0314 09:48:38.003798 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:38.454813 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv"] Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:38.561042 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-t9qwm"] Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:39.070023 4843 generic.go:334] "Generic (PLEG): container finished" podID="af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" containerID="3295c4030a7b85687715d84059a654373b7cf06c816f87b2acb9ef54d96266a6" exitCode=0 Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:39.070117 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" event={"ID":"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b","Type":"ContainerDied","Data":"3295c4030a7b85687715d84059a654373b7cf06c816f87b2acb9ef54d96266a6"} Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:39.071002 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" event={"ID":"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b","Type":"ContainerStarted","Data":"9bacd1e3930cdc99261f972c1abd27fd960284bb337bfdd8228255bc1930a2a0"} Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:39.074531 4843 generic.go:334] "Generic (PLEG): container finished" podID="be5ed09f-5725-43cc-8ae5-fc8deac795ea" containerID="d4693b83f32a63fb2df86bdd6281bbde1cf2f96696fdc87f6ee5addc316288ed" exitCode=0 Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:39.074575 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-t9qwm" event={"ID":"be5ed09f-5725-43cc-8ae5-fc8deac795ea","Type":"ContainerDied","Data":"d4693b83f32a63fb2df86bdd6281bbde1cf2f96696fdc87f6ee5addc316288ed"} Mar 14 09:48:39 crc kubenswrapper[4843]: I0314 09:48:39.074602 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-t9qwm" event={"ID":"be5ed09f-5725-43cc-8ae5-fc8deac795ea","Type":"ContainerStarted","Data":"ddd944b5b67f06f1f736fd8f172830c1a8fd3bd5edeac3a36ae12d5e4749ff52"} Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.576482 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.647861 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.744967 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx9zr\" (UniqueName: \"kubernetes.io/projected/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-kube-api-access-jx9zr\") pod \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.745141 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtrrt\" (UniqueName: \"kubernetes.io/projected/be5ed09f-5725-43cc-8ae5-fc8deac795ea-kube-api-access-xtrrt\") pod \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.745209 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be5ed09f-5725-43cc-8ae5-fc8deac795ea-operator-scripts\") pod \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\" (UID: \"be5ed09f-5725-43cc-8ae5-fc8deac795ea\") " Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.745283 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-operator-scripts\") pod \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\" (UID: \"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b\") " Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.745945 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" (UID: "af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.745978 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be5ed09f-5725-43cc-8ae5-fc8deac795ea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be5ed09f-5725-43cc-8ae5-fc8deac795ea" (UID: "be5ed09f-5725-43cc-8ae5-fc8deac795ea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.750323 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-kube-api-access-jx9zr" (OuterVolumeSpecName: "kube-api-access-jx9zr") pod "af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" (UID: "af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b"). InnerVolumeSpecName "kube-api-access-jx9zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.764046 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be5ed09f-5725-43cc-8ae5-fc8deac795ea-kube-api-access-xtrrt" (OuterVolumeSpecName: "kube-api-access-xtrrt") pod "be5ed09f-5725-43cc-8ae5-fc8deac795ea" (UID: "be5ed09f-5725-43cc-8ae5-fc8deac795ea"). InnerVolumeSpecName "kube-api-access-xtrrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.847191 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx9zr\" (UniqueName: \"kubernetes.io/projected/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-kube-api-access-jx9zr\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.847228 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtrrt\" (UniqueName: \"kubernetes.io/projected/be5ed09f-5725-43cc-8ae5-fc8deac795ea-kube-api-access-xtrrt\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.847240 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be5ed09f-5725-43cc-8ae5-fc8deac795ea-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:40 crc kubenswrapper[4843]: I0314 09:48:40.847250 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:41 crc kubenswrapper[4843]: I0314 09:48:41.091771 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" Mar 14 09:48:41 crc kubenswrapper[4843]: I0314 09:48:41.091778 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv" event={"ID":"af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b","Type":"ContainerDied","Data":"9bacd1e3930cdc99261f972c1abd27fd960284bb337bfdd8228255bc1930a2a0"} Mar 14 09:48:41 crc kubenswrapper[4843]: I0314 09:48:41.092189 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bacd1e3930cdc99261f972c1abd27fd960284bb337bfdd8228255bc1930a2a0" Mar 14 09:48:41 crc kubenswrapper[4843]: I0314 09:48:41.093728 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-t9qwm" event={"ID":"be5ed09f-5725-43cc-8ae5-fc8deac795ea","Type":"ContainerDied","Data":"ddd944b5b67f06f1f736fd8f172830c1a8fd3bd5edeac3a36ae12d5e4749ff52"} Mar 14 09:48:41 crc kubenswrapper[4843]: I0314 09:48:41.093759 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddd944b5b67f06f1f736fd8f172830c1a8fd3bd5edeac3a36ae12d5e4749ff52" Mar 14 09:48:41 crc kubenswrapper[4843]: I0314 09:48:41.093804 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-t9qwm" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.015218 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh"] Mar 14 09:48:43 crc kubenswrapper[4843]: E0314 09:48:43.015599 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" containerName="mariadb-account-create-update" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.015614 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" containerName="mariadb-account-create-update" Mar 14 09:48:43 crc kubenswrapper[4843]: E0314 09:48:43.015637 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5ed09f-5725-43cc-8ae5-fc8deac795ea" containerName="mariadb-database-create" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.015646 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5ed09f-5725-43cc-8ae5-fc8deac795ea" containerName="mariadb-database-create" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.015892 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" containerName="mariadb-account-create-update" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.015917 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5ed09f-5725-43cc-8ae5-fc8deac795ea" containerName="mariadb-database-create" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.016695 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.018520 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-6thfd" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.018912 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.028727 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh"] Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.081499 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfrch\" (UniqueName: \"kubernetes.io/projected/4c1501db-d274-4770-98ed-701dca80f711-kube-api-access-qfrch\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.081921 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-config-data\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.081992 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-db-sync-config-data\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.082019 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.183568 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfrch\" (UniqueName: \"kubernetes.io/projected/4c1501db-d274-4770-98ed-701dca80f711-kube-api-access-qfrch\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.183653 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-config-data\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.183702 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-db-sync-config-data\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.183723 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.197465 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-config-data\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.198747 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.199056 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-db-sync-config-data\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.208854 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfrch\" (UniqueName: \"kubernetes.io/projected/4c1501db-d274-4770-98ed-701dca80f711-kube-api-access-qfrch\") pod \"watcher-kuttl-db-sync-vfjdh\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.344650 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:43 crc kubenswrapper[4843]: I0314 09:48:43.828036 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh"] Mar 14 09:48:43 crc kubenswrapper[4843]: W0314 09:48:43.850178 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c1501db_d274_4770_98ed_701dca80f711.slice/crio-cc0de02ab1984c8da7a1bbff60f0534ac0823110b08e7643afc8e739bd5c08c7 WatchSource:0}: Error finding container cc0de02ab1984c8da7a1bbff60f0534ac0823110b08e7643afc8e739bd5c08c7: Status 404 returned error can't find the container with id cc0de02ab1984c8da7a1bbff60f0534ac0823110b08e7643afc8e739bd5c08c7 Mar 14 09:48:44 crc kubenswrapper[4843]: I0314 09:48:44.117319 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" event={"ID":"4c1501db-d274-4770-98ed-701dca80f711","Type":"ContainerStarted","Data":"6eda1fb53b7eff4793ea91205dc643197aebfb1a3e38508b7edc79e87efa5cd0"} Mar 14 09:48:44 crc kubenswrapper[4843]: I0314 09:48:44.117605 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" event={"ID":"4c1501db-d274-4770-98ed-701dca80f711","Type":"ContainerStarted","Data":"cc0de02ab1984c8da7a1bbff60f0534ac0823110b08e7643afc8e739bd5c08c7"} Mar 14 09:48:44 crc kubenswrapper[4843]: I0314 09:48:44.136894 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" podStartSLOduration=2.136863319 podStartE2EDuration="2.136863319s" podCreationTimestamp="2026-03-14 09:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:44.135929937 +0000 UTC m=+2291.448541065" watchObservedRunningTime="2026-03-14 09:48:44.136863319 +0000 UTC m=+2291.449474447" Mar 14 09:48:47 crc kubenswrapper[4843]: I0314 09:48:47.143128 4843 generic.go:334] "Generic (PLEG): container finished" podID="4c1501db-d274-4770-98ed-701dca80f711" containerID="6eda1fb53b7eff4793ea91205dc643197aebfb1a3e38508b7edc79e87efa5cd0" exitCode=0 Mar 14 09:48:47 crc kubenswrapper[4843]: I0314 09:48:47.143215 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" event={"ID":"4c1501db-d274-4770-98ed-701dca80f711","Type":"ContainerDied","Data":"6eda1fb53b7eff4793ea91205dc643197aebfb1a3e38508b7edc79e87efa5cd0"} Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.454028 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.567942 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfrch\" (UniqueName: \"kubernetes.io/projected/4c1501db-d274-4770-98ed-701dca80f711-kube-api-access-qfrch\") pod \"4c1501db-d274-4770-98ed-701dca80f711\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.568063 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-db-sync-config-data\") pod \"4c1501db-d274-4770-98ed-701dca80f711\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.568115 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-combined-ca-bundle\") pod \"4c1501db-d274-4770-98ed-701dca80f711\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.568147 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-config-data\") pod \"4c1501db-d274-4770-98ed-701dca80f711\" (UID: \"4c1501db-d274-4770-98ed-701dca80f711\") " Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.573573 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c1501db-d274-4770-98ed-701dca80f711-kube-api-access-qfrch" (OuterVolumeSpecName: "kube-api-access-qfrch") pod "4c1501db-d274-4770-98ed-701dca80f711" (UID: "4c1501db-d274-4770-98ed-701dca80f711"). InnerVolumeSpecName "kube-api-access-qfrch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.579986 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4c1501db-d274-4770-98ed-701dca80f711" (UID: "4c1501db-d274-4770-98ed-701dca80f711"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.601431 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c1501db-d274-4770-98ed-701dca80f711" (UID: "4c1501db-d274-4770-98ed-701dca80f711"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.629430 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-config-data" (OuterVolumeSpecName: "config-data") pod "4c1501db-d274-4770-98ed-701dca80f711" (UID: "4c1501db-d274-4770-98ed-701dca80f711"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.674485 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfrch\" (UniqueName: \"kubernetes.io/projected/4c1501db-d274-4770-98ed-701dca80f711-kube-api-access-qfrch\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.674531 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.674543 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:48 crc kubenswrapper[4843]: I0314 09:48:48.674555 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c1501db-d274-4770-98ed-701dca80f711-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.163188 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" event={"ID":"4c1501db-d274-4770-98ed-701dca80f711","Type":"ContainerDied","Data":"cc0de02ab1984c8da7a1bbff60f0534ac0823110b08e7643afc8e739bd5c08c7"} Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.163245 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc0de02ab1984c8da7a1bbff60f0534ac0823110b08e7643afc8e739bd5c08c7" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.163381 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.340000 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:48:49 crc kubenswrapper[4843]: E0314 09:48:49.340371 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.503842 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:49 crc kubenswrapper[4843]: E0314 09:48:49.504306 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c1501db-d274-4770-98ed-701dca80f711" containerName="watcher-kuttl-db-sync" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.504324 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c1501db-d274-4770-98ed-701dca80f711" containerName="watcher-kuttl-db-sync" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.504492 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c1501db-d274-4770-98ed-701dca80f711" containerName="watcher-kuttl-db-sync" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.505481 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.516921 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.517313 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.517522 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.517658 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-6thfd" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.533036 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.534498 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.541353 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.559763 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593383 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-logs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593468 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593494 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593530 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593589 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593627 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h5j7\" (UniqueName: \"kubernetes.io/projected/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-kube-api-access-5h5j7\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.593679 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.610532 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.668383 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.669911 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.674591 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.688627 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.701897 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.703197 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.703357 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69658b2e-8db1-470a-82c1-891c2e75de0c-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.703445 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.703573 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-logs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.703687 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.703809 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.704079 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.707512 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.707614 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djnrd\" (UniqueName: \"kubernetes.io/projected/69658b2e-8db1-470a-82c1-891c2e75de0c-kube-api-access-djnrd\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.707834 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.707965 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h5j7\" (UniqueName: \"kubernetes.io/projected/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-kube-api-access-5h5j7\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.707610 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-logs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.736136 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.736706 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.737173 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.742880 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h5j7\" (UniqueName: \"kubernetes.io/projected/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-kube-api-access-5h5j7\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.744053 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.759041 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.790374 4843 scope.go:117] "RemoveContainer" containerID="3a56a1c9e4716e7c3635615ee804b0e8548d002a37dd3e6841aca2a0b44d9b9c" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.818749 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819015 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819035 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djnrd\" (UniqueName: \"kubernetes.io/projected/69658b2e-8db1-470a-82c1-891c2e75de0c-kube-api-access-djnrd\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819068 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk4b6\" (UniqueName: \"kubernetes.io/projected/5bea3daf-04a8-4744-9f47-bf4fd43b844d-kube-api-access-sk4b6\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819142 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819170 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69658b2e-8db1-470a-82c1-891c2e75de0c-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819186 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819205 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bea3daf-04a8-4744-9f47-bf4fd43b844d-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.819229 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.823452 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.824895 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69658b2e-8db1-470a-82c1-891c2e75de0c-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.830091 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.843179 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.844835 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djnrd\" (UniqueName: \"kubernetes.io/projected/69658b2e-8db1-470a-82c1-891c2e75de0c-kube-api-access-djnrd\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.849659 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.852641 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.920786 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bea3daf-04a8-4744-9f47-bf4fd43b844d-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.920854 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.920901 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.920931 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk4b6\" (UniqueName: \"kubernetes.io/projected/5bea3daf-04a8-4744-9f47-bf4fd43b844d-kube-api-access-sk4b6\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.921510 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bea3daf-04a8-4744-9f47-bf4fd43b844d-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.925209 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.925813 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.947229 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk4b6\" (UniqueName: \"kubernetes.io/projected/5bea3daf-04a8-4744-9f47-bf4fd43b844d-kube-api-access-sk4b6\") pod \"watcher-kuttl-applier-0\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:49 crc kubenswrapper[4843]: I0314 09:48:49.994205 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:50 crc kubenswrapper[4843]: W0314 09:48:50.334849 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f815abc_4cef_4f67_86d2_c2dbe0f23d78.slice/crio-0ec9795137c20c104555dc64a0726ad933a78915268fd1dbd9d1806f614abb4a WatchSource:0}: Error finding container 0ec9795137c20c104555dc64a0726ad933a78915268fd1dbd9d1806f614abb4a: Status 404 returned error can't find the container with id 0ec9795137c20c104555dc64a0726ad933a78915268fd1dbd9d1806f614abb4a Mar 14 09:48:50 crc kubenswrapper[4843]: I0314 09:48:50.335769 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:48:50 crc kubenswrapper[4843]: I0314 09:48:50.419114 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:48:50 crc kubenswrapper[4843]: W0314 09:48:50.423347 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69658b2e_8db1_470a_82c1_891c2e75de0c.slice/crio-6cd5ca7213a51cadade4049118e5c769aa4fd6137225682c1bbdc4d24b6f202c WatchSource:0}: Error finding container 6cd5ca7213a51cadade4049118e5c769aa4fd6137225682c1bbdc4d24b6f202c: Status 404 returned error can't find the container with id 6cd5ca7213a51cadade4049118e5c769aa4fd6137225682c1bbdc4d24b6f202c Mar 14 09:48:50 crc kubenswrapper[4843]: I0314 09:48:50.560940 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.187893 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"5bea3daf-04a8-4744-9f47-bf4fd43b844d","Type":"ContainerStarted","Data":"c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.188233 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"5bea3daf-04a8-4744-9f47-bf4fd43b844d","Type":"ContainerStarted","Data":"8f533ab43f9033d92ab1be693ef3c3531aec6d2d41161ef2647c0d1d7533835e"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.189775 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"69658b2e-8db1-470a-82c1-891c2e75de0c","Type":"ContainerStarted","Data":"b2cca327e7ea80d472d19fb40e2ea17c79274c888c003485cead70d3a95d45e8"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.189810 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"69658b2e-8db1-470a-82c1-891c2e75de0c","Type":"ContainerStarted","Data":"6cd5ca7213a51cadade4049118e5c769aa4fd6137225682c1bbdc4d24b6f202c"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.192893 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"5f815abc-4cef-4f67-86d2-c2dbe0f23d78","Type":"ContainerStarted","Data":"5296c741372ab6e2d4e5bee922a44353c7b1bb23af7239b3c4c8417a977c0695"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.192939 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"5f815abc-4cef-4f67-86d2-c2dbe0f23d78","Type":"ContainerStarted","Data":"6e323f803b930f5e592f3ccccf0ec467b717d0cc6e9a76d557946f8622a52293"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.192954 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"5f815abc-4cef-4f67-86d2-c2dbe0f23d78","Type":"ContainerStarted","Data":"0ec9795137c20c104555dc64a0726ad933a78915268fd1dbd9d1806f614abb4a"} Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.193152 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.211931 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.211904678 podStartE2EDuration="2.211904678s" podCreationTimestamp="2026-03-14 09:48:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:51.204327778 +0000 UTC m=+2298.516938926" watchObservedRunningTime="2026-03-14 09:48:51.211904678 +0000 UTC m=+2298.524515806" Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.230705 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.230679796 podStartE2EDuration="2.230679796s" podCreationTimestamp="2026-03-14 09:48:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:51.225716747 +0000 UTC m=+2298.538327885" watchObservedRunningTime="2026-03-14 09:48:51.230679796 +0000 UTC m=+2298.543290934" Mar 14 09:48:51 crc kubenswrapper[4843]: I0314 09:48:51.266417 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.266391847 podStartE2EDuration="2.266391847s" podCreationTimestamp="2026-03-14 09:48:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:48:51.253106791 +0000 UTC m=+2298.565717919" watchObservedRunningTime="2026-03-14 09:48:51.266391847 +0000 UTC m=+2298.579002975" Mar 14 09:48:53 crc kubenswrapper[4843]: I0314 09:48:53.882370 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:54 crc kubenswrapper[4843]: I0314 09:48:54.824632 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:54 crc kubenswrapper[4843]: I0314 09:48:54.995849 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:48:59 crc kubenswrapper[4843]: I0314 09:48:59.824028 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:59 crc kubenswrapper[4843]: I0314 09:48:59.833739 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:48:59 crc kubenswrapper[4843]: I0314 09:48:59.852910 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:59 crc kubenswrapper[4843]: I0314 09:48:59.876106 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:48:59 crc kubenswrapper[4843]: I0314 09:48:59.995421 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:00 crc kubenswrapper[4843]: I0314 09:49:00.023792 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:00 crc kubenswrapper[4843]: I0314 09:49:00.260148 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:00 crc kubenswrapper[4843]: I0314 09:49:00.267181 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:00 crc kubenswrapper[4843]: I0314 09:49:00.340485 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:49:00 crc kubenswrapper[4843]: E0314 09:49:00.340707 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:49:00 crc kubenswrapper[4843]: I0314 09:49:00.663740 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:00 crc kubenswrapper[4843]: I0314 09:49:00.663798 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:02 crc kubenswrapper[4843]: I0314 09:49:02.724107 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:02 crc kubenswrapper[4843]: I0314 09:49:02.977390 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:03 crc kubenswrapper[4843]: I0314 09:49:03.282305 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="sg-core" containerID="cri-o://f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9" gracePeriod=30 Mar 14 09:49:03 crc kubenswrapper[4843]: I0314 09:49:03.282324 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="proxy-httpd" containerID="cri-o://1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f" gracePeriod=30 Mar 14 09:49:03 crc kubenswrapper[4843]: I0314 09:49:03.282408 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-central-agent" containerID="cri-o://168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f" gracePeriod=30 Mar 14 09:49:03 crc kubenswrapper[4843]: I0314 09:49:03.282392 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-notification-agent" containerID="cri-o://0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334" gracePeriod=30 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.033482 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.034042 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-kuttl-api-log" containerID="cri-o://6e323f803b930f5e592f3ccccf0ec467b717d0cc6e9a76d557946f8622a52293" gracePeriod=30 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.034083 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-api" containerID="cri-o://5296c741372ab6e2d4e5bee922a44353c7b1bb23af7239b3c4c8417a977c0695" gracePeriod=30 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.291264 4843 generic.go:334] "Generic (PLEG): container finished" podID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerID="6e323f803b930f5e592f3ccccf0ec467b717d0cc6e9a76d557946f8622a52293" exitCode=143 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.291364 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"5f815abc-4cef-4f67-86d2-c2dbe0f23d78","Type":"ContainerDied","Data":"6e323f803b930f5e592f3ccccf0ec467b717d0cc6e9a76d557946f8622a52293"} Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.295301 4843 generic.go:334] "Generic (PLEG): container finished" podID="24ab212e-5406-458a-97ef-a18305940668" containerID="1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f" exitCode=0 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.295339 4843 generic.go:334] "Generic (PLEG): container finished" podID="24ab212e-5406-458a-97ef-a18305940668" containerID="f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9" exitCode=2 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.295354 4843 generic.go:334] "Generic (PLEG): container finished" podID="24ab212e-5406-458a-97ef-a18305940668" containerID="168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f" exitCode=0 Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.295376 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerDied","Data":"1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f"} Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.295404 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerDied","Data":"f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9"} Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.295416 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerDied","Data":"168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f"} Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.730351 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.864559 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-config-data\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.864618 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-run-httpd\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.864643 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tczhm\" (UniqueName: \"kubernetes.io/projected/24ab212e-5406-458a-97ef-a18305940668-kube-api-access-tczhm\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.864695 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-combined-ca-bundle\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.865334 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.865423 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-scripts\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.865458 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-sg-core-conf-yaml\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.865495 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-log-httpd\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.865564 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-ceilometer-tls-certs\") pod \"24ab212e-5406-458a-97ef-a18305940668\" (UID: \"24ab212e-5406-458a-97ef-a18305940668\") " Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.865865 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.866344 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.869584 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-scripts" (OuterVolumeSpecName: "scripts") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.870084 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ab212e-5406-458a-97ef-a18305940668-kube-api-access-tczhm" (OuterVolumeSpecName: "kube-api-access-tczhm") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "kube-api-access-tczhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.906119 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.911489 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.911654 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.183:9322/\": read tcp 10.217.0.2:51154->10.217.0.183:9322: read: connection reset by peer" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.911670 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"https://10.217.0.183:9322/\": read tcp 10.217.0.2:51146->10.217.0.183:9322: read: connection reset by peer" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.970207 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tczhm\" (UniqueName: \"kubernetes.io/projected/24ab212e-5406-458a-97ef-a18305940668-kube-api-access-tczhm\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.970249 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.970260 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.970268 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24ab212e-5406-458a-97ef-a18305940668-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.970291 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:04 crc kubenswrapper[4843]: I0314 09:49:04.984998 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-config-data" (OuterVolumeSpecName: "config-data") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.056426 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24ab212e-5406-458a-97ef-a18305940668" (UID: "24ab212e-5406-458a-97ef-a18305940668"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.073331 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.073364 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ab212e-5406-458a-97ef-a18305940668-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.304668 4843 generic.go:334] "Generic (PLEG): container finished" podID="24ab212e-5406-458a-97ef-a18305940668" containerID="0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334" exitCode=0 Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.304806 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerDied","Data":"0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334"} Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.304842 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"24ab212e-5406-458a-97ef-a18305940668","Type":"ContainerDied","Data":"f04f6fb57bb1224a8793cc6f305ceef1eeaa8a5a895f4011c748ed29747af412"} Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.304864 4843 scope.go:117] "RemoveContainer" containerID="1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.305060 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.308507 4843 generic.go:334] "Generic (PLEG): container finished" podID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerID="5296c741372ab6e2d4e5bee922a44353c7b1bb23af7239b3c4c8417a977c0695" exitCode=0 Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.308551 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"5f815abc-4cef-4f67-86d2-c2dbe0f23d78","Type":"ContainerDied","Data":"5296c741372ab6e2d4e5bee922a44353c7b1bb23af7239b3c4c8417a977c0695"} Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.333501 4843 scope.go:117] "RemoveContainer" containerID="f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.351764 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.356683 4843 scope.go:117] "RemoveContainer" containerID="0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.374607 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.385625 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.386041 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="sg-core" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386062 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="sg-core" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.386088 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-notification-agent" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386098 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-notification-agent" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.386119 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-central-agent" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386127 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-central-agent" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.386142 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="proxy-httpd" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386151 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="proxy-httpd" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386393 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="sg-core" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386416 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-central-agent" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386444 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="ceilometer-notification-agent" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.386468 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ab212e-5406-458a-97ef-a18305940668" containerName="proxy-httpd" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.389200 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.392373 4843 scope.go:117] "RemoveContainer" containerID="168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.392865 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.394747 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.394931 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.421299 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.427380 4843 scope.go:117] "RemoveContainer" containerID="1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.427846 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f\": container with ID starting with 1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f not found: ID does not exist" containerID="1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.427890 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f"} err="failed to get container status \"1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f\": rpc error: code = NotFound desc = could not find container \"1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f\": container with ID starting with 1c249d96e2ebc5319bbbdd3dcac6bd596f758c84fe30a7a12273f73d14ceb87f not found: ID does not exist" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.427923 4843 scope.go:117] "RemoveContainer" containerID="f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.428332 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9\": container with ID starting with f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9 not found: ID does not exist" containerID="f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.428358 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9"} err="failed to get container status \"f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9\": rpc error: code = NotFound desc = could not find container \"f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9\": container with ID starting with f9f7a5dccc763d6e5fa806a01d8dbc11231f3a1bca6c82375bd71486b7bd26f9 not found: ID does not exist" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.428379 4843 scope.go:117] "RemoveContainer" containerID="0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.428932 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334\": container with ID starting with 0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334 not found: ID does not exist" containerID="0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.428967 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334"} err="failed to get container status \"0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334\": rpc error: code = NotFound desc = could not find container \"0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334\": container with ID starting with 0674adebbc5f7c5a4822a576743b8d2e3264b5f6442dd1bd0181ff8f0cc99334 not found: ID does not exist" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.428986 4843 scope.go:117] "RemoveContainer" containerID="168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f" Mar 14 09:49:05 crc kubenswrapper[4843]: E0314 09:49:05.429349 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f\": container with ID starting with 168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f not found: ID does not exist" containerID="168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.429376 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f"} err="failed to get container status \"168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f\": rpc error: code = NotFound desc = could not find container \"168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f\": container with ID starting with 168f4dea906173213fb999b9527150fe0e18bbea48599ee8d8a7c7d9f703451f not found: ID does not exist" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.470641 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482052 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-config-data\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482133 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482161 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482247 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-log-httpd\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482458 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-run-httpd\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482520 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-scripts\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482548 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.482716 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k97l\" (UniqueName: \"kubernetes.io/projected/1a081b6c-9f09-40b4-befc-4ff11bad82a8-kube-api-access-7k97l\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584169 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-combined-ca-bundle\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584263 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h5j7\" (UniqueName: \"kubernetes.io/projected/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-kube-api-access-5h5j7\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584324 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-public-tls-certs\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584346 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-config-data\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584445 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-custom-prometheus-ca\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584554 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-logs\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584580 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-internal-tls-certs\") pod \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\" (UID: \"5f815abc-4cef-4f67-86d2-c2dbe0f23d78\") " Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584823 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-config-data\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584870 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584895 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.584950 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-log-httpd\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585487 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-logs" (OuterVolumeSpecName: "logs") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585507 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-run-httpd\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585553 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585575 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-scripts\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585667 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k97l\" (UniqueName: \"kubernetes.io/projected/1a081b6c-9f09-40b4-befc-4ff11bad82a8-kube-api-access-7k97l\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585718 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.585791 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-log-httpd\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.586041 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-run-httpd\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.594081 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.594296 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.594435 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-scripts\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.594438 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-config-data\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.603889 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-kube-api-access-5h5j7" (OuterVolumeSpecName: "kube-api-access-5h5j7") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "kube-api-access-5h5j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.611003 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k97l\" (UniqueName: \"kubernetes.io/projected/1a081b6c-9f09-40b4-befc-4ff11bad82a8-kube-api-access-7k97l\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.620423 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.622456 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.630499 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.644421 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.659347 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.662913 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-config-data" (OuterVolumeSpecName: "config-data") pod "5f815abc-4cef-4f67-86d2-c2dbe0f23d78" (UID: "5f815abc-4cef-4f67-86d2-c2dbe0f23d78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.687136 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.687182 4843 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.687196 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.687208 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h5j7\" (UniqueName: \"kubernetes.io/projected/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-kube-api-access-5h5j7\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.687223 4843 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.687234 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f815abc-4cef-4f67-86d2-c2dbe0f23d78-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:05 crc kubenswrapper[4843]: I0314 09:49:05.720494 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.234667 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.317075 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"5f815abc-4cef-4f67-86d2-c2dbe0f23d78","Type":"ContainerDied","Data":"0ec9795137c20c104555dc64a0726ad933a78915268fd1dbd9d1806f614abb4a"} Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.317122 4843 scope.go:117] "RemoveContainer" containerID="5296c741372ab6e2d4e5bee922a44353c7b1bb23af7239b3c4c8417a977c0695" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.317206 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.321724 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerStarted","Data":"e6520729424d7254451dd5d39e2103e9e0a6db4c8cdd75561cf6cd4dd07dfbe9"} Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.339195 4843 scope.go:117] "RemoveContainer" containerID="6e323f803b930f5e592f3ccccf0ec467b717d0cc6e9a76d557946f8622a52293" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.354694 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.365153 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.373378 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:06 crc kubenswrapper[4843]: E0314 09:49:06.373955 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-kuttl-api-log" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.373979 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-kuttl-api-log" Mar 14 09:49:06 crc kubenswrapper[4843]: E0314 09:49:06.374049 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-api" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.374059 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-api" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.374384 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-kuttl-api-log" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.374437 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" containerName="watcher-api" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.375651 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.380608 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.380835 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.381155 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.385814 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.499848 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.500301 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0207b01-70a0-467a-9a3e-e6efaf799f5e-logs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.500439 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.500517 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c696w\" (UniqueName: \"kubernetes.io/projected/d0207b01-70a0-467a-9a3e-e6efaf799f5e-kube-api-access-c696w\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.500603 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.500684 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.500756 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.602954 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603024 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c696w\" (UniqueName: \"kubernetes.io/projected/d0207b01-70a0-467a-9a3e-e6efaf799f5e-kube-api-access-c696w\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603072 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603114 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603146 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603186 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603208 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0207b01-70a0-467a-9a3e-e6efaf799f5e-logs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.603940 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0207b01-70a0-467a-9a3e-e6efaf799f5e-logs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.610755 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.616039 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.616391 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.616764 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.617712 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.633248 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c696w\" (UniqueName: \"kubernetes.io/projected/d0207b01-70a0-467a-9a3e-e6efaf799f5e-kube-api-access-c696w\") pod \"watcher-kuttl-api-0\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:06 crc kubenswrapper[4843]: I0314 09:49:06.700258 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:07 crc kubenswrapper[4843]: I0314 09:49:07.183151 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:07 crc kubenswrapper[4843]: I0314 09:49:07.364207 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ab212e-5406-458a-97ef-a18305940668" path="/var/lib/kubelet/pods/24ab212e-5406-458a-97ef-a18305940668/volumes" Mar 14 09:49:07 crc kubenswrapper[4843]: I0314 09:49:07.365130 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f815abc-4cef-4f67-86d2-c2dbe0f23d78" path="/var/lib/kubelet/pods/5f815abc-4cef-4f67-86d2-c2dbe0f23d78/volumes" Mar 14 09:49:07 crc kubenswrapper[4843]: I0314 09:49:07.365832 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerStarted","Data":"8810fc272c6db2e0500ebd829142983d60fa4b552a2b8955dabf836d0791da6c"} Mar 14 09:49:07 crc kubenswrapper[4843]: I0314 09:49:07.365911 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d0207b01-70a0-467a-9a3e-e6efaf799f5e","Type":"ContainerStarted","Data":"e8c6189943d4c5dbffa7f77c448892b4cd67e659bbc8ecc83a163df2d6c569d5"} Mar 14 09:49:08 crc kubenswrapper[4843]: I0314 09:49:08.376013 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d0207b01-70a0-467a-9a3e-e6efaf799f5e","Type":"ContainerStarted","Data":"c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5"} Mar 14 09:49:08 crc kubenswrapper[4843]: I0314 09:49:08.376488 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:08 crc kubenswrapper[4843]: I0314 09:49:08.376503 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d0207b01-70a0-467a-9a3e-e6efaf799f5e","Type":"ContainerStarted","Data":"a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c"} Mar 14 09:49:08 crc kubenswrapper[4843]: I0314 09:49:08.382347 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerStarted","Data":"bfb81cff9a76d452c3cedb1e2619478e97ec9841c6a6f16dc2387aebe1e317f9"} Mar 14 09:49:08 crc kubenswrapper[4843]: I0314 09:49:08.382384 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerStarted","Data":"b83f30fdc93542c2d7333c08b27da9682859c7a974197ccca915937a87cb2ec4"} Mar 14 09:49:08 crc kubenswrapper[4843]: I0314 09:49:08.404391 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.404367007 podStartE2EDuration="2.404367007s" podCreationTimestamp="2026-03-14 09:49:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:49:08.399330188 +0000 UTC m=+2315.711941326" watchObservedRunningTime="2026-03-14 09:49:08.404367007 +0000 UTC m=+2315.716978135" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.098129 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.107456 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-vfjdh"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.140060 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher5aaf-account-delete-h5zpx"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.141064 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.160748 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher5aaf-account-delete-h5zpx"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.170782 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.171065 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" containerName="watcher-applier" containerID="cri-o://c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" gracePeriod=30 Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.210624 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.219074 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.219515 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="69658b2e-8db1-470a-82c1-891c2e75de0c" containerName="watcher-decision-engine" containerID="cri-o://b2cca327e7ea80d472d19fb40e2ea17c79274c888c003485cead70d3a95d45e8" gracePeriod=30 Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.250308 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e81dc339-60c2-4d08-bbc6-9535406b487d-operator-scripts\") pod \"watcher5aaf-account-delete-h5zpx\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.250464 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flfwp\" (UniqueName: \"kubernetes.io/projected/e81dc339-60c2-4d08-bbc6-9535406b487d-kube-api-access-flfwp\") pod \"watcher5aaf-account-delete-h5zpx\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.351486 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flfwp\" (UniqueName: \"kubernetes.io/projected/e81dc339-60c2-4d08-bbc6-9535406b487d-kube-api-access-flfwp\") pod \"watcher5aaf-account-delete-h5zpx\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.351588 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e81dc339-60c2-4d08-bbc6-9535406b487d-operator-scripts\") pod \"watcher5aaf-account-delete-h5zpx\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.352553 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e81dc339-60c2-4d08-bbc6-9535406b487d-operator-scripts\") pod \"watcher5aaf-account-delete-h5zpx\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.356304 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c1501db-d274-4770-98ed-701dca80f711" path="/var/lib/kubelet/pods/4c1501db-d274-4770-98ed-701dca80f711/volumes" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.374860 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flfwp\" (UniqueName: \"kubernetes.io/projected/e81dc339-60c2-4d08-bbc6-9535406b487d-kube-api-access-flfwp\") pod \"watcher5aaf-account-delete-h5zpx\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.389731 4843 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="watcher-kuttl-default/watcher-kuttl-api-0" secret="" err="secret \"watcher-watcher-kuttl-dockercfg-6thfd\" not found" Mar 14 09:49:09 crc kubenswrapper[4843]: E0314 09:49:09.453809 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:09 crc kubenswrapper[4843]: E0314 09:49:09.453865 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data podName:d0207b01-70a0-467a-9a3e-e6efaf799f5e nodeName:}" failed. No retries permitted until 2026-03-14 09:49:09.953849405 +0000 UTC m=+2317.266460533 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data") pod "watcher-kuttl-api-0" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e") : secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.456417 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:09 crc kubenswrapper[4843]: E0314 09:49:09.959666 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:09 crc kubenswrapper[4843]: E0314 09:49:09.960288 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data podName:d0207b01-70a0-467a-9a3e-e6efaf799f5e nodeName:}" failed. No retries permitted until 2026-03-14 09:49:10.960250816 +0000 UTC m=+2318.272861944 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data") pod "watcher-kuttl-api-0" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e") : secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:09 crc kubenswrapper[4843]: I0314 09:49:09.986939 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher5aaf-account-delete-h5zpx"] Mar 14 09:49:09 crc kubenswrapper[4843]: W0314 09:49:09.990624 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode81dc339_60c2_4d08_bbc6_9535406b487d.slice/crio-6de1878989c0f89a9b9067cb6eed023ad89c2387101a0fc95ed80ca6560b0c84 WatchSource:0}: Error finding container 6de1878989c0f89a9b9067cb6eed023ad89c2387101a0fc95ed80ca6560b0c84: Status 404 returned error can't find the container with id 6de1878989c0f89a9b9067cb6eed023ad89c2387101a0fc95ed80ca6560b0c84 Mar 14 09:49:09 crc kubenswrapper[4843]: E0314 09:49:09.997885 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:49:09 crc kubenswrapper[4843]: E0314 09:49:09.999473 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:49:10 crc kubenswrapper[4843]: E0314 09:49:10.001111 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:49:10 crc kubenswrapper[4843]: E0314 09:49:10.001159 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" containerName="watcher-applier" Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.399504 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerStarted","Data":"464703728cc0430235fad3e12c96e2524f72d61d0dde6370c5b38160d7af4956"} Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.399679 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.401568 4843 generic.go:334] "Generic (PLEG): container finished" podID="e81dc339-60c2-4d08-bbc6-9535406b487d" containerID="b0814bae3cb1837930d651cccd41bb5520013bb18641d5a98b8fcb4fdcd647b4" exitCode=0 Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.401671 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.401692 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" event={"ID":"e81dc339-60c2-4d08-bbc6-9535406b487d","Type":"ContainerDied","Data":"b0814bae3cb1837930d651cccd41bb5520013bb18641d5a98b8fcb4fdcd647b4"} Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.401731 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" event={"ID":"e81dc339-60c2-4d08-bbc6-9535406b487d","Type":"ContainerStarted","Data":"6de1878989c0f89a9b9067cb6eed023ad89c2387101a0fc95ed80ca6560b0c84"} Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.401845 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-kuttl-api-log" containerID="cri-o://c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5" gracePeriod=30 Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.401906 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" containerID="cri-o://a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c" gracePeriod=30 Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.408065 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.187:9322/\": EOF" Mar 14 09:49:10 crc kubenswrapper[4843]: I0314 09:49:10.428967 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.916238275 podStartE2EDuration="5.428948118s" podCreationTimestamp="2026-03-14 09:49:05 +0000 UTC" firstStartedPulling="2026-03-14 09:49:06.248730134 +0000 UTC m=+2313.561341262" lastFinishedPulling="2026-03-14 09:49:09.761439977 +0000 UTC m=+2317.074051105" observedRunningTime="2026-03-14 09:49:10.422343041 +0000 UTC m=+2317.734954179" watchObservedRunningTime="2026-03-14 09:49:10.428948118 +0000 UTC m=+2317.741559246" Mar 14 09:49:10 crc kubenswrapper[4843]: E0314 09:49:10.977139 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:10 crc kubenswrapper[4843]: E0314 09:49:10.977545 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data podName:d0207b01-70a0-467a-9a3e-e6efaf799f5e nodeName:}" failed. No retries permitted until 2026-03-14 09:49:12.977521844 +0000 UTC m=+2320.290132972 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data") pod "watcher-kuttl-api-0" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e") : secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.411324 4843 generic.go:334] "Generic (PLEG): container finished" podID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerID="c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5" exitCode=143 Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.411423 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d0207b01-70a0-467a-9a3e-e6efaf799f5e","Type":"ContainerDied","Data":"c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5"} Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.700445 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.789396 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.895982 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flfwp\" (UniqueName: \"kubernetes.io/projected/e81dc339-60c2-4d08-bbc6-9535406b487d-kube-api-access-flfwp\") pod \"e81dc339-60c2-4d08-bbc6-9535406b487d\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.896102 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e81dc339-60c2-4d08-bbc6-9535406b487d-operator-scripts\") pod \"e81dc339-60c2-4d08-bbc6-9535406b487d\" (UID: \"e81dc339-60c2-4d08-bbc6-9535406b487d\") " Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.897357 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e81dc339-60c2-4d08-bbc6-9535406b487d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e81dc339-60c2-4d08-bbc6-9535406b487d" (UID: "e81dc339-60c2-4d08-bbc6-9535406b487d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.911395 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81dc339-60c2-4d08-bbc6-9535406b487d-kube-api-access-flfwp" (OuterVolumeSpecName: "kube-api-access-flfwp") pod "e81dc339-60c2-4d08-bbc6-9535406b487d" (UID: "e81dc339-60c2-4d08-bbc6-9535406b487d"). InnerVolumeSpecName "kube-api-access-flfwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.991483 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.997741 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flfwp\" (UniqueName: \"kubernetes.io/projected/e81dc339-60c2-4d08-bbc6-9535406b487d-kube-api-access-flfwp\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:11 crc kubenswrapper[4843]: I0314 09:49:11.997770 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e81dc339-60c2-4d08-bbc6-9535406b487d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.099014 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk4b6\" (UniqueName: \"kubernetes.io/projected/5bea3daf-04a8-4744-9f47-bf4fd43b844d-kube-api-access-sk4b6\") pod \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.099106 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-combined-ca-bundle\") pod \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.099140 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-config-data\") pod \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.099216 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bea3daf-04a8-4744-9f47-bf4fd43b844d-logs\") pod \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\" (UID: \"5bea3daf-04a8-4744-9f47-bf4fd43b844d\") " Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.099568 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bea3daf-04a8-4744-9f47-bf4fd43b844d-logs" (OuterVolumeSpecName: "logs") pod "5bea3daf-04a8-4744-9f47-bf4fd43b844d" (UID: "5bea3daf-04a8-4744-9f47-bf4fd43b844d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.099647 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bea3daf-04a8-4744-9f47-bf4fd43b844d-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.101852 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bea3daf-04a8-4744-9f47-bf4fd43b844d-kube-api-access-sk4b6" (OuterVolumeSpecName: "kube-api-access-sk4b6") pod "5bea3daf-04a8-4744-9f47-bf4fd43b844d" (UID: "5bea3daf-04a8-4744-9f47-bf4fd43b844d"). InnerVolumeSpecName "kube-api-access-sk4b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.125772 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bea3daf-04a8-4744-9f47-bf4fd43b844d" (UID: "5bea3daf-04a8-4744-9f47-bf4fd43b844d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.144868 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-config-data" (OuterVolumeSpecName: "config-data") pod "5bea3daf-04a8-4744-9f47-bf4fd43b844d" (UID: "5bea3daf-04a8-4744-9f47-bf4fd43b844d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.201013 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk4b6\" (UniqueName: \"kubernetes.io/projected/5bea3daf-04a8-4744-9f47-bf4fd43b844d-kube-api-access-sk4b6\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.201064 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.201076 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bea3daf-04a8-4744-9f47-bf4fd43b844d-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.330515 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.419945 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.419963 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher5aaf-account-delete-h5zpx" event={"ID":"e81dc339-60c2-4d08-bbc6-9535406b487d","Type":"ContainerDied","Data":"6de1878989c0f89a9b9067cb6eed023ad89c2387101a0fc95ed80ca6560b0c84"} Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.420019 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6de1878989c0f89a9b9067cb6eed023ad89c2387101a0fc95ed80ca6560b0c84" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.421585 4843 generic.go:334] "Generic (PLEG): container finished" podID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" exitCode=0 Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.421629 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"5bea3daf-04a8-4744-9f47-bf4fd43b844d","Type":"ContainerDied","Data":"c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e"} Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.421650 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.421673 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"5bea3daf-04a8-4744-9f47-bf4fd43b844d","Type":"ContainerDied","Data":"8f533ab43f9033d92ab1be693ef3c3531aec6d2d41161ef2647c0d1d7533835e"} Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.421695 4843 scope.go:117] "RemoveContainer" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.422170 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="sg-core" containerID="cri-o://bfb81cff9a76d452c3cedb1e2619478e97ec9841c6a6f16dc2387aebe1e317f9" gracePeriod=30 Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.422188 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="proxy-httpd" containerID="cri-o://464703728cc0430235fad3e12c96e2524f72d61d0dde6370c5b38160d7af4956" gracePeriod=30 Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.422170 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-central-agent" containerID="cri-o://8810fc272c6db2e0500ebd829142983d60fa4b552a2b8955dabf836d0791da6c" gracePeriod=30 Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.422229 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-notification-agent" containerID="cri-o://b83f30fdc93542c2d7333c08b27da9682859c7a974197ccca915937a87cb2ec4" gracePeriod=30 Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.466414 4843 scope.go:117] "RemoveContainer" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" Mar 14 09:49:12 crc kubenswrapper[4843]: E0314 09:49:12.473595 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e\": container with ID starting with c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e not found: ID does not exist" containerID="c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.473636 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e"} err="failed to get container status \"c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e\": rpc error: code = NotFound desc = could not find container \"c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e\": container with ID starting with c3d037aa944fc8cf93fe01575cd7f6002118af4a573b141f04d64324c8c07b0e not found: ID does not exist" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.495726 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.524455 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.529430 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.187:9322/\": read tcp 10.217.0.2:38554->10.217.0.187:9322: read: connection reset by peer" Mar 14 09:49:12 crc kubenswrapper[4843]: I0314 09:49:12.529933 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.187:9322/\": dial tcp 10.217.0.187:9322: connect: connection refused" Mar 14 09:49:12 crc kubenswrapper[4843]: E0314 09:49:12.956194 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a081b6c_9f09_40b4_befc_4ff11bad82a8.slice/crio-464703728cc0430235fad3e12c96e2524f72d61d0dde6370c5b38160d7af4956.scope\": RecentStats: unable to find data in memory cache]" Mar 14 09:49:13 crc kubenswrapper[4843]: E0314 09:49:13.018466 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-api-config-data: secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:13 crc kubenswrapper[4843]: E0314 09:49:13.018541 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data podName:d0207b01-70a0-467a-9a3e-e6efaf799f5e nodeName:}" failed. No retries permitted until 2026-03-14 09:49:17.018522916 +0000 UTC m=+2324.331134044 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data") pod "watcher-kuttl-api-0" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e") : secret "watcher-kuttl-api-config-data" not found Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.198083 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322204 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-public-tls-certs\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322319 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c696w\" (UniqueName: \"kubernetes.io/projected/d0207b01-70a0-467a-9a3e-e6efaf799f5e-kube-api-access-c696w\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322414 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-internal-tls-certs\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322442 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-combined-ca-bundle\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322464 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-custom-prometheus-ca\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322526 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.322549 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0207b01-70a0-467a-9a3e-e6efaf799f5e-logs\") pod \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\" (UID: \"d0207b01-70a0-467a-9a3e-e6efaf799f5e\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.323162 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0207b01-70a0-467a-9a3e-e6efaf799f5e-logs" (OuterVolumeSpecName: "logs") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.337562 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0207b01-70a0-467a-9a3e-e6efaf799f5e-kube-api-access-c696w" (OuterVolumeSpecName: "kube-api-access-c696w") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "kube-api-access-c696w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.367508 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" path="/var/lib/kubelet/pods/5bea3daf-04a8-4744-9f47-bf4fd43b844d/volumes" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.382342 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.404126 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.406939 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data" (OuterVolumeSpecName: "config-data") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.408734 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.410901 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d0207b01-70a0-467a-9a3e-e6efaf799f5e" (UID: "d0207b01-70a0-467a-9a3e-e6efaf799f5e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424654 4843 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424685 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c696w\" (UniqueName: \"kubernetes.io/projected/d0207b01-70a0-467a-9a3e-e6efaf799f5e-kube-api-access-c696w\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424696 4843 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424706 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424714 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424724 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0207b01-70a0-467a-9a3e-e6efaf799f5e-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.424735 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0207b01-70a0-467a-9a3e-e6efaf799f5e-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.432214 4843 generic.go:334] "Generic (PLEG): container finished" podID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerID="a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c" exitCode=0 Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.432300 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.432307 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d0207b01-70a0-467a-9a3e-e6efaf799f5e","Type":"ContainerDied","Data":"a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c"} Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.432336 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d0207b01-70a0-467a-9a3e-e6efaf799f5e","Type":"ContainerDied","Data":"e8c6189943d4c5dbffa7f77c448892b4cd67e659bbc8ecc83a163df2d6c569d5"} Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.432357 4843 scope.go:117] "RemoveContainer" containerID="a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440562 4843 generic.go:334] "Generic (PLEG): container finished" podID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerID="464703728cc0430235fad3e12c96e2524f72d61d0dde6370c5b38160d7af4956" exitCode=0 Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440717 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerDied","Data":"464703728cc0430235fad3e12c96e2524f72d61d0dde6370c5b38160d7af4956"} Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440800 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerDied","Data":"bfb81cff9a76d452c3cedb1e2619478e97ec9841c6a6f16dc2387aebe1e317f9"} Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440740 4843 generic.go:334] "Generic (PLEG): container finished" podID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerID="bfb81cff9a76d452c3cedb1e2619478e97ec9841c6a6f16dc2387aebe1e317f9" exitCode=2 Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440825 4843 generic.go:334] "Generic (PLEG): container finished" podID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerID="b83f30fdc93542c2d7333c08b27da9682859c7a974197ccca915937a87cb2ec4" exitCode=0 Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440862 4843 generic.go:334] "Generic (PLEG): container finished" podID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerID="8810fc272c6db2e0500ebd829142983d60fa4b552a2b8955dabf836d0791da6c" exitCode=0 Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440851 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerDied","Data":"b83f30fdc93542c2d7333c08b27da9682859c7a974197ccca915937a87cb2ec4"} Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.440900 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerDied","Data":"8810fc272c6db2e0500ebd829142983d60fa4b552a2b8955dabf836d0791da6c"} Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.455383 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.468101 4843 scope.go:117] "RemoveContainer" containerID="c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.471085 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.477489 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.493491 4843 scope.go:117] "RemoveContainer" containerID="a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c" Mar 14 09:49:13 crc kubenswrapper[4843]: E0314 09:49:13.497831 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c\": container with ID starting with a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c not found: ID does not exist" containerID="a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.497869 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c"} err="failed to get container status \"a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c\": rpc error: code = NotFound desc = could not find container \"a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c\": container with ID starting with a269f9b9e378a031e907c86fdb60736648630dd2e462ede2791b0b3c1d65e75c not found: ID does not exist" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.497893 4843 scope.go:117] "RemoveContainer" containerID="c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5" Mar 14 09:49:13 crc kubenswrapper[4843]: E0314 09:49:13.498599 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5\": container with ID starting with c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5 not found: ID does not exist" containerID="c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.498646 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5"} err="failed to get container status \"c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5\": rpc error: code = NotFound desc = could not find container \"c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5\": container with ID starting with c28298a1eebe5117d456ce2dcd9dda3ed45fcb1873cc33fdedeb215fb05102c5 not found: ID does not exist" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526118 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k97l\" (UniqueName: \"kubernetes.io/projected/1a081b6c-9f09-40b4-befc-4ff11bad82a8-kube-api-access-7k97l\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526219 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-scripts\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526267 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-combined-ca-bundle\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526319 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-sg-core-conf-yaml\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526352 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-ceilometer-tls-certs\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526370 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-log-httpd\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526392 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-config-data\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526437 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-run-httpd\") pod \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\" (UID: \"1a081b6c-9f09-40b4-befc-4ff11bad82a8\") " Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526898 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.526916 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.530409 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-scripts" (OuterVolumeSpecName: "scripts") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.530566 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a081b6c-9f09-40b4-befc-4ff11bad82a8-kube-api-access-7k97l" (OuterVolumeSpecName: "kube-api-access-7k97l") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "kube-api-access-7k97l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.551059 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.567251 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.596673 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.624298 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-config-data" (OuterVolumeSpecName: "config-data") pod "1a081b6c-9f09-40b4-befc-4ff11bad82a8" (UID: "1a081b6c-9f09-40b4-befc-4ff11bad82a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628320 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628427 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628575 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628596 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628607 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a081b6c-9f09-40b4-befc-4ff11bad82a8-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628619 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k97l\" (UniqueName: \"kubernetes.io/projected/1a081b6c-9f09-40b4-befc-4ff11bad82a8-kube-api-access-7k97l\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628631 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:13 crc kubenswrapper[4843]: I0314 09:49:13.628642 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a081b6c-9f09-40b4-befc-4ff11bad82a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.189324 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher5aaf-account-delete-h5zpx"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.199499 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.209166 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher5aaf-account-delete-h5zpx"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.215789 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-t9qwm"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.221973 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-5aaf-account-create-update-v5dbv"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.228594 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-t9qwm"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.468660 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1a081b6c-9f09-40b4-befc-4ff11bad82a8","Type":"ContainerDied","Data":"e6520729424d7254451dd5d39e2103e9e0a6db4c8cdd75561cf6cd4dd07dfbe9"} Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.468977 4843 scope.go:117] "RemoveContainer" containerID="464703728cc0430235fad3e12c96e2524f72d61d0dde6370c5b38160d7af4956" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.468732 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.493451 4843 scope.go:117] "RemoveContainer" containerID="bfb81cff9a76d452c3cedb1e2619478e97ec9841c6a6f16dc2387aebe1e317f9" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.514424 4843 scope.go:117] "RemoveContainer" containerID="b83f30fdc93542c2d7333c08b27da9682859c7a974197ccca915937a87cb2ec4" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.520223 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.544315 4843 scope.go:117] "RemoveContainer" containerID="8810fc272c6db2e0500ebd829142983d60fa4b552a2b8955dabf836d0791da6c" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.544613 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.577726 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578346 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" containerName="watcher-applier" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578363 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" containerName="watcher-applier" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578377 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578384 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578396 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-kuttl-api-log" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578422 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-kuttl-api-log" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578433 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="proxy-httpd" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578439 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="proxy-httpd" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578450 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="sg-core" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578456 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="sg-core" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578477 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-central-agent" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578483 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-central-agent" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578498 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81dc339-60c2-4d08-bbc6-9535406b487d" containerName="mariadb-account-delete" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578506 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81dc339-60c2-4d08-bbc6-9535406b487d" containerName="mariadb-account-delete" Mar 14 09:49:14 crc kubenswrapper[4843]: E0314 09:49:14.578516 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-notification-agent" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578524 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-notification-agent" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578675 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-api" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578693 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="sg-core" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578701 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81dc339-60c2-4d08-bbc6-9535406b487d" containerName="mariadb-account-delete" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578711 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bea3daf-04a8-4744-9f47-bf4fd43b844d" containerName="watcher-applier" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578717 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="proxy-httpd" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578729 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" containerName="watcher-kuttl-api-log" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578737 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-central-agent" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.578748 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" containerName="ceilometer-notification-agent" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.581024 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.585316 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.585403 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.585533 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.588176 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647049 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-run-httpd\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647181 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wln2g\" (UniqueName: \"kubernetes.io/projected/a431bf38-ad51-4b73-928d-bfe4e76d827f-kube-api-access-wln2g\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647216 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-config-data\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647308 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647343 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647383 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-scripts\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647405 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.647433 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-log-httpd\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749124 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749171 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-scripts\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749196 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749225 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-log-httpd\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749298 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-run-httpd\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749336 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wln2g\" (UniqueName: \"kubernetes.io/projected/a431bf38-ad51-4b73-928d-bfe4e76d827f-kube-api-access-wln2g\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749352 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-config-data\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.749382 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.750449 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-log-httpd\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.750921 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-run-httpd\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.754975 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.754982 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.756056 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-scripts\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.756412 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-config-data\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.757317 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.782259 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wln2g\" (UniqueName: \"kubernetes.io/projected/a431bf38-ad51-4b73-928d-bfe4e76d827f-kube-api-access-wln2g\") pod \"ceilometer-0\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:14 crc kubenswrapper[4843]: I0314 09:49:14.909743 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.339235 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:49:15 crc kubenswrapper[4843]: E0314 09:49:15.339834 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.348431 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a081b6c-9f09-40b4-befc-4ff11bad82a8" path="/var/lib/kubelet/pods/1a081b6c-9f09-40b4-befc-4ff11bad82a8/volumes" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.349359 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b" path="/var/lib/kubelet/pods/af50dd62-c1c2-4c8e-aed0-4e6bc1cc0e8b/volumes" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.349927 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be5ed09f-5725-43cc-8ae5-fc8deac795ea" path="/var/lib/kubelet/pods/be5ed09f-5725-43cc-8ae5-fc8deac795ea/volumes" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.351259 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0207b01-70a0-467a-9a3e-e6efaf799f5e" path="/var/lib/kubelet/pods/d0207b01-70a0-467a-9a3e-e6efaf799f5e/volumes" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.351817 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e81dc339-60c2-4d08-bbc6-9535406b487d" path="/var/lib/kubelet/pods/e81dc339-60c2-4d08-bbc6-9535406b487d/volumes" Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.403547 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:15 crc kubenswrapper[4843]: I0314 09:49:15.480123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerStarted","Data":"e5f036dda11bc102fed98b7ba7c9e33dd8a47a14eea2e67057ad01f283ce6cc3"} Mar 14 09:49:16 crc kubenswrapper[4843]: I0314 09:49:16.498876 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerStarted","Data":"49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732"} Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.509169 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerStarted","Data":"5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709"} Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.509485 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerStarted","Data":"4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456"} Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.514881 4843 generic.go:334] "Generic (PLEG): container finished" podID="69658b2e-8db1-470a-82c1-891c2e75de0c" containerID="b2cca327e7ea80d472d19fb40e2ea17c79274c888c003485cead70d3a95d45e8" exitCode=0 Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.514911 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"69658b2e-8db1-470a-82c1-891c2e75de0c","Type":"ContainerDied","Data":"b2cca327e7ea80d472d19fb40e2ea17c79274c888c003485cead70d3a95d45e8"} Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.767759 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.800555 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-custom-prometheus-ca\") pod \"69658b2e-8db1-470a-82c1-891c2e75de0c\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.800675 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-config-data\") pod \"69658b2e-8db1-470a-82c1-891c2e75de0c\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.800715 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-combined-ca-bundle\") pod \"69658b2e-8db1-470a-82c1-891c2e75de0c\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.800773 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69658b2e-8db1-470a-82c1-891c2e75de0c-logs\") pod \"69658b2e-8db1-470a-82c1-891c2e75de0c\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.800849 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djnrd\" (UniqueName: \"kubernetes.io/projected/69658b2e-8db1-470a-82c1-891c2e75de0c-kube-api-access-djnrd\") pod \"69658b2e-8db1-470a-82c1-891c2e75de0c\" (UID: \"69658b2e-8db1-470a-82c1-891c2e75de0c\") " Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.804915 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69658b2e-8db1-470a-82c1-891c2e75de0c-logs" (OuterVolumeSpecName: "logs") pod "69658b2e-8db1-470a-82c1-891c2e75de0c" (UID: "69658b2e-8db1-470a-82c1-891c2e75de0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.820877 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69658b2e-8db1-470a-82c1-891c2e75de0c-kube-api-access-djnrd" (OuterVolumeSpecName: "kube-api-access-djnrd") pod "69658b2e-8db1-470a-82c1-891c2e75de0c" (UID: "69658b2e-8db1-470a-82c1-891c2e75de0c"). InnerVolumeSpecName "kube-api-access-djnrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.840967 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69658b2e-8db1-470a-82c1-891c2e75de0c" (UID: "69658b2e-8db1-470a-82c1-891c2e75de0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.845432 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-config-data" (OuterVolumeSpecName: "config-data") pod "69658b2e-8db1-470a-82c1-891c2e75de0c" (UID: "69658b2e-8db1-470a-82c1-891c2e75de0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.852599 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "69658b2e-8db1-470a-82c1-891c2e75de0c" (UID: "69658b2e-8db1-470a-82c1-891c2e75de0c"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.902954 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.902984 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.902993 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69658b2e-8db1-470a-82c1-891c2e75de0c-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.903010 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djnrd\" (UniqueName: \"kubernetes.io/projected/69658b2e-8db1-470a-82c1-891c2e75de0c-kube-api-access-djnrd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:17 crc kubenswrapper[4843]: I0314 09:49:17.903019 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/69658b2e-8db1-470a-82c1-891c2e75de0c-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:18 crc kubenswrapper[4843]: I0314 09:49:18.525220 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"69658b2e-8db1-470a-82c1-891c2e75de0c","Type":"ContainerDied","Data":"6cd5ca7213a51cadade4049118e5c769aa4fd6137225682c1bbdc4d24b6f202c"} Mar 14 09:49:18 crc kubenswrapper[4843]: I0314 09:49:18.525297 4843 scope.go:117] "RemoveContainer" containerID="b2cca327e7ea80d472d19fb40e2ea17c79274c888c003485cead70d3a95d45e8" Mar 14 09:49:18 crc kubenswrapper[4843]: I0314 09:49:18.525339 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:18 crc kubenswrapper[4843]: I0314 09:49:18.585617 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:18 crc kubenswrapper[4843]: I0314 09:49:18.594574 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.349924 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69658b2e-8db1-470a-82c1-891c2e75de0c" path="/var/lib/kubelet/pods/69658b2e-8db1-470a-82c1-891c2e75de0c/volumes" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.534325 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerStarted","Data":"29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996"} Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.535389 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.564436 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.290104981 podStartE2EDuration="5.564417302s" podCreationTimestamp="2026-03-14 09:49:14 +0000 UTC" firstStartedPulling="2026-03-14 09:49:15.414477619 +0000 UTC m=+2322.727088737" lastFinishedPulling="2026-03-14 09:49:18.68878993 +0000 UTC m=+2326.001401058" observedRunningTime="2026-03-14 09:49:19.558735716 +0000 UTC m=+2326.871346844" watchObservedRunningTime="2026-03-14 09:49:19.564417302 +0000 UTC m=+2326.877028430" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.734512 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-5tng5"] Mar 14 09:49:19 crc kubenswrapper[4843]: E0314 09:49:19.734835 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69658b2e-8db1-470a-82c1-891c2e75de0c" containerName="watcher-decision-engine" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.734851 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="69658b2e-8db1-470a-82c1-891c2e75de0c" containerName="watcher-decision-engine" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.734997 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="69658b2e-8db1-470a-82c1-891c2e75de0c" containerName="watcher-decision-engine" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.735494 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.751550 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5tng5"] Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.840498 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5"] Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.841762 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.844225 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.845109 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh4fc\" (UniqueName: \"kubernetes.io/projected/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-kube-api-access-zh4fc\") pod \"watcher-db-create-5tng5\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.845179 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-operator-scripts\") pod \"watcher-db-create-5tng5\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.848760 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5"] Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.947360 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-operator-scripts\") pod \"watcher-c17a-account-create-update-jtsj5\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.947424 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw6cg\" (UniqueName: \"kubernetes.io/projected/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-kube-api-access-qw6cg\") pod \"watcher-c17a-account-create-update-jtsj5\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.947497 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh4fc\" (UniqueName: \"kubernetes.io/projected/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-kube-api-access-zh4fc\") pod \"watcher-db-create-5tng5\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.947543 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-operator-scripts\") pod \"watcher-db-create-5tng5\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.948492 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-operator-scripts\") pod \"watcher-db-create-5tng5\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:19 crc kubenswrapper[4843]: I0314 09:49:19.974193 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh4fc\" (UniqueName: \"kubernetes.io/projected/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-kube-api-access-zh4fc\") pod \"watcher-db-create-5tng5\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.049033 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-operator-scripts\") pod \"watcher-c17a-account-create-update-jtsj5\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.049079 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw6cg\" (UniqueName: \"kubernetes.io/projected/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-kube-api-access-qw6cg\") pod \"watcher-c17a-account-create-update-jtsj5\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.049838 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-operator-scripts\") pod \"watcher-c17a-account-create-update-jtsj5\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.051842 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.065762 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw6cg\" (UniqueName: \"kubernetes.io/projected/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-kube-api-access-qw6cg\") pod \"watcher-c17a-account-create-update-jtsj5\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.156097 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.553966 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5tng5"] Mar 14 09:49:20 crc kubenswrapper[4843]: I0314 09:49:20.635313 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5"] Mar 14 09:49:20 crc kubenswrapper[4843]: W0314 09:49:20.639012 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c780c8c_e7e7_41b8_adc6_0c8726e1d7f9.slice/crio-e67b2ad6252d2ff858c4704c141b9ceb2a0807918e9e6aaabda674c71bf5dbac WatchSource:0}: Error finding container e67b2ad6252d2ff858c4704c141b9ceb2a0807918e9e6aaabda674c71bf5dbac: Status 404 returned error can't find the container with id e67b2ad6252d2ff858c4704c141b9ceb2a0807918e9e6aaabda674c71bf5dbac Mar 14 09:49:21 crc kubenswrapper[4843]: I0314 09:49:21.560303 4843 generic.go:334] "Generic (PLEG): container finished" podID="2f55d604-7f52-420f-ac0c-7d3e68a2bffc" containerID="b63f9b1db597dc2dad973419315a3bc0625aeeb62ed1e2d467bf0ac2f49f913e" exitCode=0 Mar 14 09:49:21 crc kubenswrapper[4843]: I0314 09:49:21.560396 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5tng5" event={"ID":"2f55d604-7f52-420f-ac0c-7d3e68a2bffc","Type":"ContainerDied","Data":"b63f9b1db597dc2dad973419315a3bc0625aeeb62ed1e2d467bf0ac2f49f913e"} Mar 14 09:49:21 crc kubenswrapper[4843]: I0314 09:49:21.560430 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5tng5" event={"ID":"2f55d604-7f52-420f-ac0c-7d3e68a2bffc","Type":"ContainerStarted","Data":"b855d42cbe4f4ad7239309e6243c38637c9a2fac11c97dc814489b6ebac52b61"} Mar 14 09:49:21 crc kubenswrapper[4843]: I0314 09:49:21.561875 4843 generic.go:334] "Generic (PLEG): container finished" podID="8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" containerID="ad2888b2c4d2a3bfb91ed2f3be3d9eca1437462b443ff21fb53a06decb96c392" exitCode=0 Mar 14 09:49:21 crc kubenswrapper[4843]: I0314 09:49:21.562942 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" event={"ID":"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9","Type":"ContainerDied","Data":"ad2888b2c4d2a3bfb91ed2f3be3d9eca1437462b443ff21fb53a06decb96c392"} Mar 14 09:49:21 crc kubenswrapper[4843]: I0314 09:49:21.563000 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" event={"ID":"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9","Type":"ContainerStarted","Data":"e67b2ad6252d2ff858c4704c141b9ceb2a0807918e9e6aaabda674c71bf5dbac"} Mar 14 09:49:22 crc kubenswrapper[4843]: I0314 09:49:22.968849 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.002126 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-operator-scripts\") pod \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.002301 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh4fc\" (UniqueName: \"kubernetes.io/projected/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-kube-api-access-zh4fc\") pod \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\" (UID: \"2f55d604-7f52-420f-ac0c-7d3e68a2bffc\") " Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.002795 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f55d604-7f52-420f-ac0c-7d3e68a2bffc" (UID: "2f55d604-7f52-420f-ac0c-7d3e68a2bffc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.003001 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.010104 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-kube-api-access-zh4fc" (OuterVolumeSpecName: "kube-api-access-zh4fc") pod "2f55d604-7f52-420f-ac0c-7d3e68a2bffc" (UID: "2f55d604-7f52-420f-ac0c-7d3e68a2bffc"). InnerVolumeSpecName "kube-api-access-zh4fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.021990 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.104065 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw6cg\" (UniqueName: \"kubernetes.io/projected/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-kube-api-access-qw6cg\") pod \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.104121 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-operator-scripts\") pod \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\" (UID: \"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9\") " Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.104609 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh4fc\" (UniqueName: \"kubernetes.io/projected/2f55d604-7f52-420f-ac0c-7d3e68a2bffc-kube-api-access-zh4fc\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.104705 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" (UID: "8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.115456 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-kube-api-access-qw6cg" (OuterVolumeSpecName: "kube-api-access-qw6cg") pod "8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" (UID: "8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9"). InnerVolumeSpecName "kube-api-access-qw6cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.207040 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw6cg\" (UniqueName: \"kubernetes.io/projected/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-kube-api-access-qw6cg\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.207068 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.577313 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.577332 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5" event={"ID":"8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9","Type":"ContainerDied","Data":"e67b2ad6252d2ff858c4704c141b9ceb2a0807918e9e6aaabda674c71bf5dbac"} Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.577386 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e67b2ad6252d2ff858c4704c141b9ceb2a0807918e9e6aaabda674c71bf5dbac" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.579035 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-5tng5" event={"ID":"2f55d604-7f52-420f-ac0c-7d3e68a2bffc","Type":"ContainerDied","Data":"b855d42cbe4f4ad7239309e6243c38637c9a2fac11c97dc814489b6ebac52b61"} Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.579066 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b855d42cbe4f4ad7239309e6243c38637c9a2fac11c97dc814489b6ebac52b61" Mar 14 09:49:23 crc kubenswrapper[4843]: I0314 09:49:23.579183 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-5tng5" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.184058 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-x29w7"] Mar 14 09:49:25 crc kubenswrapper[4843]: E0314 09:49:25.184981 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" containerName="mariadb-account-create-update" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.184997 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" containerName="mariadb-account-create-update" Mar 14 09:49:25 crc kubenswrapper[4843]: E0314 09:49:25.185013 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f55d604-7f52-420f-ac0c-7d3e68a2bffc" containerName="mariadb-database-create" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.185020 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f55d604-7f52-420f-ac0c-7d3e68a2bffc" containerName="mariadb-database-create" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.185208 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f55d604-7f52-420f-ac0c-7d3e68a2bffc" containerName="mariadb-database-create" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.185235 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" containerName="mariadb-account-create-update" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.185888 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.187846 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.189371 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-n6xj7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.204659 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-x29w7"] Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.240378 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-config-data\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.240673 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-db-sync-config-data\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.240847 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46827\" (UniqueName: \"kubernetes.io/projected/448d3cda-6240-4734-b55a-65a8d33a69f6-kube-api-access-46827\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.240981 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.342229 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-db-sync-config-data\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.342366 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46827\" (UniqueName: \"kubernetes.io/projected/448d3cda-6240-4734-b55a-65a8d33a69f6-kube-api-access-46827\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.342400 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.342440 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-config-data\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.352917 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-db-sync-config-data\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.359891 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.361959 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-config-data\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.381867 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46827\" (UniqueName: \"kubernetes.io/projected/448d3cda-6240-4734-b55a-65a8d33a69f6-kube-api-access-46827\") pod \"watcher-kuttl-db-sync-x29w7\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:25 crc kubenswrapper[4843]: I0314 09:49:25.503547 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:26 crc kubenswrapper[4843]: I0314 09:49:26.025943 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-x29w7"] Mar 14 09:49:26 crc kubenswrapper[4843]: W0314 09:49:26.027709 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod448d3cda_6240_4734_b55a_65a8d33a69f6.slice/crio-079105e3148400fc57d126a920e01533f19c82bff7c69a3019bba39d44ef248e WatchSource:0}: Error finding container 079105e3148400fc57d126a920e01533f19c82bff7c69a3019bba39d44ef248e: Status 404 returned error can't find the container with id 079105e3148400fc57d126a920e01533f19c82bff7c69a3019bba39d44ef248e Mar 14 09:49:26 crc kubenswrapper[4843]: I0314 09:49:26.339374 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:49:26 crc kubenswrapper[4843]: E0314 09:49:26.339865 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:49:26 crc kubenswrapper[4843]: I0314 09:49:26.603088 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" event={"ID":"448d3cda-6240-4734-b55a-65a8d33a69f6","Type":"ContainerStarted","Data":"cb97cad51ce45cd869003032d25e4fd917ed6e7ba21bff54560f41ed7b1ce607"} Mar 14 09:49:26 crc kubenswrapper[4843]: I0314 09:49:26.603134 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" event={"ID":"448d3cda-6240-4734-b55a-65a8d33a69f6","Type":"ContainerStarted","Data":"079105e3148400fc57d126a920e01533f19c82bff7c69a3019bba39d44ef248e"} Mar 14 09:49:26 crc kubenswrapper[4843]: I0314 09:49:26.623237 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" podStartSLOduration=1.6232159830000001 podStartE2EDuration="1.623215983s" podCreationTimestamp="2026-03-14 09:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:49:26.616308659 +0000 UTC m=+2333.928919807" watchObservedRunningTime="2026-03-14 09:49:26.623215983 +0000 UTC m=+2333.935827111" Mar 14 09:49:28 crc kubenswrapper[4843]: I0314 09:49:28.620186 4843 generic.go:334] "Generic (PLEG): container finished" podID="448d3cda-6240-4734-b55a-65a8d33a69f6" containerID="cb97cad51ce45cd869003032d25e4fd917ed6e7ba21bff54560f41ed7b1ce607" exitCode=0 Mar 14 09:49:28 crc kubenswrapper[4843]: I0314 09:49:28.620285 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" event={"ID":"448d3cda-6240-4734-b55a-65a8d33a69f6","Type":"ContainerDied","Data":"cb97cad51ce45cd869003032d25e4fd917ed6e7ba21bff54560f41ed7b1ce607"} Mar 14 09:49:29 crc kubenswrapper[4843]: I0314 09:49:29.913109 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.023462 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-combined-ca-bundle\") pod \"448d3cda-6240-4734-b55a-65a8d33a69f6\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.023602 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-config-data\") pod \"448d3cda-6240-4734-b55a-65a8d33a69f6\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.023748 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46827\" (UniqueName: \"kubernetes.io/projected/448d3cda-6240-4734-b55a-65a8d33a69f6-kube-api-access-46827\") pod \"448d3cda-6240-4734-b55a-65a8d33a69f6\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.023797 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-db-sync-config-data\") pod \"448d3cda-6240-4734-b55a-65a8d33a69f6\" (UID: \"448d3cda-6240-4734-b55a-65a8d33a69f6\") " Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.028331 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "448d3cda-6240-4734-b55a-65a8d33a69f6" (UID: "448d3cda-6240-4734-b55a-65a8d33a69f6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.028832 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/448d3cda-6240-4734-b55a-65a8d33a69f6-kube-api-access-46827" (OuterVolumeSpecName: "kube-api-access-46827") pod "448d3cda-6240-4734-b55a-65a8d33a69f6" (UID: "448d3cda-6240-4734-b55a-65a8d33a69f6"). InnerVolumeSpecName "kube-api-access-46827". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.053591 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "448d3cda-6240-4734-b55a-65a8d33a69f6" (UID: "448d3cda-6240-4734-b55a-65a8d33a69f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.065434 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-config-data" (OuterVolumeSpecName: "config-data") pod "448d3cda-6240-4734-b55a-65a8d33a69f6" (UID: "448d3cda-6240-4734-b55a-65a8d33a69f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.125946 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46827\" (UniqueName: \"kubernetes.io/projected/448d3cda-6240-4734-b55a-65a8d33a69f6-kube-api-access-46827\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.126220 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.126350 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.126426 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/448d3cda-6240-4734-b55a-65a8d33a69f6-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.637192 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" event={"ID":"448d3cda-6240-4734-b55a-65a8d33a69f6","Type":"ContainerDied","Data":"079105e3148400fc57d126a920e01533f19c82bff7c69a3019bba39d44ef248e"} Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.637237 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-x29w7" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.637248 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="079105e3148400fc57d126a920e01533f19c82bff7c69a3019bba39d44ef248e" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.923381 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:30 crc kubenswrapper[4843]: E0314 09:49:30.923819 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="448d3cda-6240-4734-b55a-65a8d33a69f6" containerName="watcher-kuttl-db-sync" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.923840 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="448d3cda-6240-4734-b55a-65a8d33a69f6" containerName="watcher-kuttl-db-sync" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.924113 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="448d3cda-6240-4734-b55a-65a8d33a69f6" containerName="watcher-kuttl-db-sync" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.925211 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.927900 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.928783 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-n6xj7" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.929718 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.931045 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Mar 14 09:49:30 crc kubenswrapper[4843]: I0314 09:49:30.934675 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.032744 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.033704 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.037406 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.041972 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.042028 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.042080 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.042153 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.042179 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snjlm\" (UniqueName: \"kubernetes.io/projected/2ae4d592-7c50-4199-b933-40099a89832d-kube-api-access-snjlm\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.042392 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.042518 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae4d592-7c50-4199-b933-40099a89832d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.045322 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.046549 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.048232 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.054821 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.064045 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144181 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/781f9edc-16e2-44e0-9260-1a8a61fc4950-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144243 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144266 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snjlm\" (UniqueName: \"kubernetes.io/projected/2ae4d592-7c50-4199-b933-40099a89832d-kube-api-access-snjlm\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144321 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9mnz\" (UniqueName: \"kubernetes.io/projected/96005028-fce4-421d-8a81-85bf81e1742a-kube-api-access-z9mnz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144345 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144393 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144420 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144571 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96005028-fce4-421d-8a81-85bf81e1742a-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144638 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae4d592-7c50-4199-b933-40099a89832d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144717 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144735 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144774 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144859 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgbft\" (UniqueName: \"kubernetes.io/projected/781f9edc-16e2-44e0-9260-1a8a61fc4950-kube-api-access-lgbft\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144886 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.144927 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.145017 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae4d592-7c50-4199-b933-40099a89832d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.145043 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.148507 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.148688 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.148696 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.158816 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.161803 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.161950 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snjlm\" (UniqueName: \"kubernetes.io/projected/2ae4d592-7c50-4199-b933-40099a89832d-kube-api-access-snjlm\") pod \"watcher-kuttl-api-0\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.241569 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246177 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246234 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96005028-fce4-421d-8a81-85bf81e1742a-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246281 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246323 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgbft\" (UniqueName: \"kubernetes.io/projected/781f9edc-16e2-44e0-9260-1a8a61fc4950-kube-api-access-lgbft\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246349 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246368 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246391 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/781f9edc-16e2-44e0-9260-1a8a61fc4950-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246424 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9mnz\" (UniqueName: \"kubernetes.io/projected/96005028-fce4-421d-8a81-85bf81e1742a-kube-api-access-z9mnz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.246441 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.247322 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96005028-fce4-421d-8a81-85bf81e1742a-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.247688 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/781f9edc-16e2-44e0-9260-1a8a61fc4950-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.251949 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.252770 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.254594 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.254907 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.255015 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.267565 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9mnz\" (UniqueName: \"kubernetes.io/projected/96005028-fce4-421d-8a81-85bf81e1742a-kube-api-access-z9mnz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.268497 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgbft\" (UniqueName: \"kubernetes.io/projected/781f9edc-16e2-44e0-9260-1a8a61fc4950-kube-api-access-lgbft\") pod \"watcher-kuttl-applier-0\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.350116 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.367769 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.698808 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: W0314 09:49:31.706108 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ae4d592_7c50_4199_b933_40099a89832d.slice/crio-974d5c78bc11f427c1a3f2d5419c2a3178fb8265f222a3aaed76f7f8d9ddc8c0 WatchSource:0}: Error finding container 974d5c78bc11f427c1a3f2d5419c2a3178fb8265f222a3aaed76f7f8d9ddc8c0: Status 404 returned error can't find the container with id 974d5c78bc11f427c1a3f2d5419c2a3178fb8265f222a3aaed76f7f8d9ddc8c0 Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.836098 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: I0314 09:49:31.877427 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:31 crc kubenswrapper[4843]: W0314 09:49:31.889374 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96005028_fce4_421d_8a81_85bf81e1742a.slice/crio-b97ea8c5d8e99b54b7cdb7cc65e39eac66526f790fb1c72122be7bfe351aa076 WatchSource:0}: Error finding container b97ea8c5d8e99b54b7cdb7cc65e39eac66526f790fb1c72122be7bfe351aa076: Status 404 returned error can't find the container with id b97ea8c5d8e99b54b7cdb7cc65e39eac66526f790fb1c72122be7bfe351aa076 Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.658003 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"96005028-fce4-421d-8a81-85bf81e1742a","Type":"ContainerStarted","Data":"59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.658064 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"96005028-fce4-421d-8a81-85bf81e1742a","Type":"ContainerStarted","Data":"b97ea8c5d8e99b54b7cdb7cc65e39eac66526f790fb1c72122be7bfe351aa076"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.661510 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"2ae4d592-7c50-4199-b933-40099a89832d","Type":"ContainerStarted","Data":"bd02424abe7d682a36e22ac815021887580e3be4a905e7e179e343c7815d4a88"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.661568 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"2ae4d592-7c50-4199-b933-40099a89832d","Type":"ContainerStarted","Data":"14dfef585e74df4bd8cb39675afb6b50972f6c6d844225d3abbf6e5a6d2cbb26"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.661582 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"2ae4d592-7c50-4199-b933-40099a89832d","Type":"ContainerStarted","Data":"974d5c78bc11f427c1a3f2d5419c2a3178fb8265f222a3aaed76f7f8d9ddc8c0"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.661910 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.667877 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"781f9edc-16e2-44e0-9260-1a8a61fc4950","Type":"ContainerStarted","Data":"91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.667935 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"781f9edc-16e2-44e0-9260-1a8a61fc4950","Type":"ContainerStarted","Data":"0d59a920f6810e331c38b3e9359152519349751fe6ed6d8e3c05116030126b88"} Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.704597 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=1.704573356 podStartE2EDuration="1.704573356s" podCreationTimestamp="2026-03-14 09:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:49:32.676248841 +0000 UTC m=+2339.988859979" watchObservedRunningTime="2026-03-14 09:49:32.704573356 +0000 UTC m=+2340.017184484" Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.722985 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=1.722970884 podStartE2EDuration="1.722970884s" podCreationTimestamp="2026-03-14 09:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:49:32.722013081 +0000 UTC m=+2340.034624209" watchObservedRunningTime="2026-03-14 09:49:32.722970884 +0000 UTC m=+2340.035582012" Mar 14 09:49:32 crc kubenswrapper[4843]: I0314 09:49:32.755967 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.755944661 podStartE2EDuration="2.755944661s" podCreationTimestamp="2026-03-14 09:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:49:32.744472456 +0000 UTC m=+2340.057083604" watchObservedRunningTime="2026-03-14 09:49:32.755944661 +0000 UTC m=+2340.068555799" Mar 14 09:49:35 crc kubenswrapper[4843]: I0314 09:49:35.079901 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:36 crc kubenswrapper[4843]: I0314 09:49:36.242658 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:36 crc kubenswrapper[4843]: I0314 09:49:36.368557 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:40 crc kubenswrapper[4843]: I0314 09:49:40.339477 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:49:40 crc kubenswrapper[4843]: E0314 09:49:40.340658 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.241822 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.251452 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.350447 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.369401 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.377088 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.400938 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.736897 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.747575 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.768402 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:49:41 crc kubenswrapper[4843]: I0314 09:49:41.777556 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:43 crc kubenswrapper[4843]: I0314 09:49:43.969401 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:43 crc kubenswrapper[4843]: I0314 09:49:43.970926 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-central-agent" containerID="cri-o://49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732" gracePeriod=30 Mar 14 09:49:43 crc kubenswrapper[4843]: I0314 09:49:43.971040 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="proxy-httpd" containerID="cri-o://29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996" gracePeriod=30 Mar 14 09:49:43 crc kubenswrapper[4843]: I0314 09:49:43.971039 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="sg-core" containerID="cri-o://5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709" gracePeriod=30 Mar 14 09:49:43 crc kubenswrapper[4843]: I0314 09:49:43.971040 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-notification-agent" containerID="cri-o://4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456" gracePeriod=30 Mar 14 09:49:43 crc kubenswrapper[4843]: I0314 09:49:43.981717 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.189:3000/\": read tcp 10.217.0.2:34644->10.217.0.189:3000: read: connection reset by peer" Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.758305 4843 generic.go:334] "Generic (PLEG): container finished" podID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerID="29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996" exitCode=0 Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.758346 4843 generic.go:334] "Generic (PLEG): container finished" podID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerID="5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709" exitCode=2 Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.758357 4843 generic.go:334] "Generic (PLEG): container finished" podID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerID="49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732" exitCode=0 Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.758358 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerDied","Data":"29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996"} Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.758388 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerDied","Data":"5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709"} Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.758399 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerDied","Data":"49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732"} Mar 14 09:49:44 crc kubenswrapper[4843]: I0314 09:49:44.911567 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.189:3000/\": dial tcp 10.217.0.189:3000: connect: connection refused" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.747075 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.784257 4843 generic.go:334] "Generic (PLEG): container finished" podID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerID="4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456" exitCode=0 Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.784325 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerDied","Data":"4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456"} Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.784360 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"a431bf38-ad51-4b73-928d-bfe4e76d827f","Type":"ContainerDied","Data":"e5f036dda11bc102fed98b7ba7c9e33dd8a47a14eea2e67057ad01f283ce6cc3"} Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.784381 4843 scope.go:117] "RemoveContainer" containerID="29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.784877 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.801165 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-ceilometer-tls-certs\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.801225 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-scripts\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.801252 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-sg-core-conf-yaml\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.801996 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-run-httpd\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.802094 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-log-httpd\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.802175 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wln2g\" (UniqueName: \"kubernetes.io/projected/a431bf38-ad51-4b73-928d-bfe4e76d827f-kube-api-access-wln2g\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.802197 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-config-data\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.802309 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-combined-ca-bundle\") pod \"a431bf38-ad51-4b73-928d-bfe4e76d827f\" (UID: \"a431bf38-ad51-4b73-928d-bfe4e76d827f\") " Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.805025 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.806751 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.812514 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a431bf38-ad51-4b73-928d-bfe4e76d827f-kube-api-access-wln2g" (OuterVolumeSpecName: "kube-api-access-wln2g") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "kube-api-access-wln2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.824091 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-scripts" (OuterVolumeSpecName: "scripts") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.824853 4843 scope.go:117] "RemoveContainer" containerID="5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.829887 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.850415 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.863632 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.887882 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-config-data" (OuterVolumeSpecName: "config-data") pod "a431bf38-ad51-4b73-928d-bfe4e76d827f" (UID: "a431bf38-ad51-4b73-928d-bfe4e76d827f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.888922 4843 scope.go:117] "RemoveContainer" containerID="4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904235 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wln2g\" (UniqueName: \"kubernetes.io/projected/a431bf38-ad51-4b73-928d-bfe4e76d827f-kube-api-access-wln2g\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904282 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904293 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904301 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904309 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904317 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a431bf38-ad51-4b73-928d-bfe4e76d827f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904325 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.904334 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a431bf38-ad51-4b73-928d-bfe4e76d827f-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.916081 4843 scope.go:117] "RemoveContainer" containerID="49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.934948 4843 scope.go:117] "RemoveContainer" containerID="29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996" Mar 14 09:49:45 crc kubenswrapper[4843]: E0314 09:49:45.935436 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996\": container with ID starting with 29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996 not found: ID does not exist" containerID="29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.935487 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996"} err="failed to get container status \"29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996\": rpc error: code = NotFound desc = could not find container \"29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996\": container with ID starting with 29f4d0afd41b6f5acddbd1a738db563f788a830932f5e57272c5f11566250996 not found: ID does not exist" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.935518 4843 scope.go:117] "RemoveContainer" containerID="5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709" Mar 14 09:49:45 crc kubenswrapper[4843]: E0314 09:49:45.935841 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709\": container with ID starting with 5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709 not found: ID does not exist" containerID="5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.935877 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709"} err="failed to get container status \"5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709\": rpc error: code = NotFound desc = could not find container \"5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709\": container with ID starting with 5ef749c8a0c34eea67c429547aab88600c33209dba3da010b15d3c8f19df5709 not found: ID does not exist" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.935903 4843 scope.go:117] "RemoveContainer" containerID="4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456" Mar 14 09:49:45 crc kubenswrapper[4843]: E0314 09:49:45.936307 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456\": container with ID starting with 4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456 not found: ID does not exist" containerID="4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.936376 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456"} err="failed to get container status \"4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456\": rpc error: code = NotFound desc = could not find container \"4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456\": container with ID starting with 4568cf2b972d5acbf4368a609d9c541263df9d315390822cdf7614f06dc12456 not found: ID does not exist" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.936415 4843 scope.go:117] "RemoveContainer" containerID="49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732" Mar 14 09:49:45 crc kubenswrapper[4843]: E0314 09:49:45.936726 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732\": container with ID starting with 49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732 not found: ID does not exist" containerID="49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732" Mar 14 09:49:45 crc kubenswrapper[4843]: I0314 09:49:45.936768 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732"} err="failed to get container status \"49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732\": rpc error: code = NotFound desc = could not find container \"49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732\": container with ID starting with 49687399b6a6e305cfab33bac7972dd23c3b6fe1f377ad22b315824bcfbc3732 not found: ID does not exist" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.124453 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.135414 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.147712 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:46 crc kubenswrapper[4843]: E0314 09:49:46.148111 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="sg-core" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148131 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="sg-core" Mar 14 09:49:46 crc kubenswrapper[4843]: E0314 09:49:46.148165 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-central-agent" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148174 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-central-agent" Mar 14 09:49:46 crc kubenswrapper[4843]: E0314 09:49:46.148187 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-notification-agent" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148195 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-notification-agent" Mar 14 09:49:46 crc kubenswrapper[4843]: E0314 09:49:46.148208 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="proxy-httpd" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148218 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="proxy-httpd" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148406 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-central-agent" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148426 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="ceilometer-notification-agent" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148445 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="sg-core" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.148460 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" containerName="proxy-httpd" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.150249 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.153191 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.153932 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.154050 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.153957 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208262 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-config-data\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208382 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208405 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208423 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfvhm\" (UniqueName: \"kubernetes.io/projected/28ca6ab9-a606-4631-8714-dbca98f6cfcd-kube-api-access-gfvhm\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208476 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208518 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-run-httpd\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208543 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-log-httpd\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.208591 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-scripts\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.309934 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.309986 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.310016 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfvhm\" (UniqueName: \"kubernetes.io/projected/28ca6ab9-a606-4631-8714-dbca98f6cfcd-kube-api-access-gfvhm\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.310069 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.310112 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-run-httpd\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.310133 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-log-httpd\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.310180 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-scripts\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.310204 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-config-data\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.311191 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-log-httpd\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.311331 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-run-httpd\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.314632 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.315460 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.315654 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.316143 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-config-data\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.316691 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-scripts\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.342851 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfvhm\" (UniqueName: \"kubernetes.io/projected/28ca6ab9-a606-4631-8714-dbca98f6cfcd-kube-api-access-gfvhm\") pod \"ceilometer-0\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:46 crc kubenswrapper[4843]: I0314 09:49:46.473564 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:47 crc kubenswrapper[4843]: I0314 09:49:47.038120 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:49:47 crc kubenswrapper[4843]: I0314 09:49:47.349026 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a431bf38-ad51-4b73-928d-bfe4e76d827f" path="/var/lib/kubelet/pods/a431bf38-ad51-4b73-928d-bfe4e76d827f/volumes" Mar 14 09:49:47 crc kubenswrapper[4843]: I0314 09:49:47.803144 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerStarted","Data":"0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa"} Mar 14 09:49:47 crc kubenswrapper[4843]: I0314 09:49:47.803456 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerStarted","Data":"81a9075c3f68009b6ac980491dbaddafcacdd4976e61a58c22a4d9a6c6aa4fa6"} Mar 14 09:49:48 crc kubenswrapper[4843]: I0314 09:49:48.813834 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerStarted","Data":"f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7"} Mar 14 09:49:49 crc kubenswrapper[4843]: I0314 09:49:49.824870 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerStarted","Data":"d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f"} Mar 14 09:49:51 crc kubenswrapper[4843]: I0314 09:49:51.342916 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:49:51 crc kubenswrapper[4843]: E0314 09:49:51.343401 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:49:51 crc kubenswrapper[4843]: I0314 09:49:51.842496 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerStarted","Data":"13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75"} Mar 14 09:49:51 crc kubenswrapper[4843]: I0314 09:49:51.843637 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:49:51 crc kubenswrapper[4843]: I0314 09:49:51.870016 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.325012033 podStartE2EDuration="5.869993645s" podCreationTimestamp="2026-03-14 09:49:46 +0000 UTC" firstStartedPulling="2026-03-14 09:49:47.046810514 +0000 UTC m=+2354.359421642" lastFinishedPulling="2026-03-14 09:49:50.591792086 +0000 UTC m=+2357.904403254" observedRunningTime="2026-03-14 09:49:51.868591672 +0000 UTC m=+2359.181202810" watchObservedRunningTime="2026-03-14 09:49:51.869993645 +0000 UTC m=+2359.182604793" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.204586 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.206662 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="96005028-fce4-421d-8a81-85bf81e1742a" containerName="watcher-decision-engine" containerID="cri-o://59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c" gracePeriod=30 Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.228518 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.229049 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/memcached-0" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" containerName="memcached" containerID="cri-o://fc89194c3dae86995ed493f4a9bcad9fe005794878d0df571dbc3e997178bcb8" gracePeriod=30 Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.299315 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.299591 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-kuttl-api-log" containerID="cri-o://14dfef585e74df4bd8cb39675afb6b50972f6c6d844225d3abbf6e5a6d2cbb26" gracePeriod=30 Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.299665 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-api" containerID="cri-o://bd02424abe7d682a36e22ac815021887580e3be4a905e7e179e343c7815d4a88" gracePeriod=30 Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.315674 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.315992 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="781f9edc-16e2-44e0-9260-1a8a61fc4950" containerName="watcher-applier" containerID="cri-o://91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" gracePeriod=30 Mar 14 09:49:56 crc kubenswrapper[4843]: E0314 09:49:56.371220 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.374555 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-gw9wj"] Mar 14 09:49:56 crc kubenswrapper[4843]: E0314 09:49:56.382837 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.384449 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: E0314 09:49:56.390714 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:49:56 crc kubenswrapper[4843]: E0314 09:49:56.390838 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="781f9edc-16e2-44e0-9260-1a8a61fc4950" containerName="watcher-applier" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.392178 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"osp-secret" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.392400 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-memcached-mtls" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.410629 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-gw9wj"] Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495209 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-fernet-keys\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495261 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-credential-keys\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495314 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-cert-memcached-mtls\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495380 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-combined-ca-bundle\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495451 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-scripts\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495479 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-config-data\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.495580 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4zfm\" (UniqueName: \"kubernetes.io/projected/ffab3465-469a-45ae-8633-0d22589ae869-kube-api-access-p4zfm\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.596947 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-combined-ca-bundle\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.597232 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-scripts\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.597256 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-config-data\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.597325 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4zfm\" (UniqueName: \"kubernetes.io/projected/ffab3465-469a-45ae-8633-0d22589ae869-kube-api-access-p4zfm\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.597380 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-fernet-keys\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.597397 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-credential-keys\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.597423 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-cert-memcached-mtls\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.603326 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-credential-keys\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.603331 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-cert-memcached-mtls\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.603433 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-scripts\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.604189 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-fernet-keys\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.605450 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-combined-ca-bundle\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.611526 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-config-data\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.616646 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4zfm\" (UniqueName: \"kubernetes.io/projected/ffab3465-469a-45ae-8633-0d22589ae869-kube-api-access-p4zfm\") pod \"keystone-bootstrap-gw9wj\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.750621 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.906088 4843 generic.go:334] "Generic (PLEG): container finished" podID="2ae4d592-7c50-4199-b933-40099a89832d" containerID="14dfef585e74df4bd8cb39675afb6b50972f6c6d844225d3abbf6e5a6d2cbb26" exitCode=143 Mar 14 09:49:56 crc kubenswrapper[4843]: I0314 09:49:56.906158 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"2ae4d592-7c50-4199-b933-40099a89832d","Type":"ContainerDied","Data":"14dfef585e74df4bd8cb39675afb6b50972f6c6d844225d3abbf6e5a6d2cbb26"} Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.054561 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-gw9wj"] Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.917884 4843 generic.go:334] "Generic (PLEG): container finished" podID="8c74421b-0ff7-450b-abcc-8e374ff891ac" containerID="fc89194c3dae86995ed493f4a9bcad9fe005794878d0df571dbc3e997178bcb8" exitCode=0 Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.917960 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"8c74421b-0ff7-450b-abcc-8e374ff891ac","Type":"ContainerDied","Data":"fc89194c3dae86995ed493f4a9bcad9fe005794878d0df571dbc3e997178bcb8"} Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.921092 4843 generic.go:334] "Generic (PLEG): container finished" podID="2ae4d592-7c50-4199-b933-40099a89832d" containerID="bd02424abe7d682a36e22ac815021887580e3be4a905e7e179e343c7815d4a88" exitCode=0 Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.921167 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"2ae4d592-7c50-4199-b933-40099a89832d","Type":"ContainerDied","Data":"bd02424abe7d682a36e22ac815021887580e3be4a905e7e179e343c7815d4a88"} Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.921211 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"2ae4d592-7c50-4199-b933-40099a89832d","Type":"ContainerDied","Data":"974d5c78bc11f427c1a3f2d5419c2a3178fb8265f222a3aaed76f7f8d9ddc8c0"} Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.921226 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974d5c78bc11f427c1a3f2d5419c2a3178fb8265f222a3aaed76f7f8d9ddc8c0" Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.924211 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" event={"ID":"ffab3465-469a-45ae-8633-0d22589ae869","Type":"ContainerStarted","Data":"39518908e826d8a0f2bf3e7f8ca5097f2aa3bddfc661fced13a318202bf7f790"} Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.924244 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" event={"ID":"ffab3465-469a-45ae-8633-0d22589ae869","Type":"ContainerStarted","Data":"b21057d64bbadf46a77b0eb3a7e7f98002d06a9e310ad4763f82b69433f346bd"} Mar 14 09:49:57 crc kubenswrapper[4843]: I0314 09:49:57.950235 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" podStartSLOduration=1.950212091 podStartE2EDuration="1.950212091s" podCreationTimestamp="2026-03-14 09:49:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:49:57.949616696 +0000 UTC m=+2365.262227824" watchObservedRunningTime="2026-03-14 09:49:57.950212091 +0000 UTC m=+2365.262823219" Mar 14 09:49:58 crc kubenswrapper[4843]: E0314 09:49:58.019167 4843 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.162:43624->38.102.83.162:34757: write tcp 38.102.83.162:43624->38.102.83.162:34757: write: broken pipe Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.038818 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.121006 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-internal-tls-certs\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.121362 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-custom-prometheus-ca\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.122103 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-public-tls-certs\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.122232 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae4d592-7c50-4199-b933-40099a89832d-logs\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.122425 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-config-data\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.122534 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-combined-ca-bundle\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.122625 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snjlm\" (UniqueName: \"kubernetes.io/projected/2ae4d592-7c50-4199-b933-40099a89832d-kube-api-access-snjlm\") pod \"2ae4d592-7c50-4199-b933-40099a89832d\" (UID: \"2ae4d592-7c50-4199-b933-40099a89832d\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.122922 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ae4d592-7c50-4199-b933-40099a89832d-logs" (OuterVolumeSpecName: "logs") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.123455 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae4d592-7c50-4199-b933-40099a89832d-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.131512 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae4d592-7c50-4199-b933-40099a89832d-kube-api-access-snjlm" (OuterVolumeSpecName: "kube-api-access-snjlm") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "kube-api-access-snjlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.149148 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.158706 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.172654 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-config-data" (OuterVolumeSpecName: "config-data") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.178359 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.194848 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2ae4d592-7c50-4199-b933-40099a89832d" (UID: "2ae4d592-7c50-4199-b933-40099a89832d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.226461 4843 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.226504 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.226516 4843 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.226526 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.226538 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae4d592-7c50-4199-b933-40099a89832d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.226549 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snjlm\" (UniqueName: \"kubernetes.io/projected/2ae4d592-7c50-4199-b933-40099a89832d-kube-api-access-snjlm\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.227909 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.327081 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-kolla-config\") pod \"8c74421b-0ff7-450b-abcc-8e374ff891ac\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.327383 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm8c9\" (UniqueName: \"kubernetes.io/projected/8c74421b-0ff7-450b-abcc-8e374ff891ac-kube-api-access-sm8c9\") pod \"8c74421b-0ff7-450b-abcc-8e374ff891ac\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.327518 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-config-data\") pod \"8c74421b-0ff7-450b-abcc-8e374ff891ac\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.327609 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-memcached-tls-certs\") pod \"8c74421b-0ff7-450b-abcc-8e374ff891ac\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.327686 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-combined-ca-bundle\") pod \"8c74421b-0ff7-450b-abcc-8e374ff891ac\" (UID: \"8c74421b-0ff7-450b-abcc-8e374ff891ac\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.327959 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "8c74421b-0ff7-450b-abcc-8e374ff891ac" (UID: "8c74421b-0ff7-450b-abcc-8e374ff891ac"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.328062 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-config-data" (OuterVolumeSpecName: "config-data") pod "8c74421b-0ff7-450b-abcc-8e374ff891ac" (UID: "8c74421b-0ff7-450b-abcc-8e374ff891ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.328263 4843 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.328347 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c74421b-0ff7-450b-abcc-8e374ff891ac-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.330333 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c74421b-0ff7-450b-abcc-8e374ff891ac-kube-api-access-sm8c9" (OuterVolumeSpecName: "kube-api-access-sm8c9") pod "8c74421b-0ff7-450b-abcc-8e374ff891ac" (UID: "8c74421b-0ff7-450b-abcc-8e374ff891ac"). InnerVolumeSpecName "kube-api-access-sm8c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.363310 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c74421b-0ff7-450b-abcc-8e374ff891ac" (UID: "8c74421b-0ff7-450b-abcc-8e374ff891ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.405998 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "8c74421b-0ff7-450b-abcc-8e374ff891ac" (UID: "8c74421b-0ff7-450b-abcc-8e374ff891ac"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.430091 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm8c9\" (UniqueName: \"kubernetes.io/projected/8c74421b-0ff7-450b-abcc-8e374ff891ac-kube-api-access-sm8c9\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.430121 4843 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.430131 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c74421b-0ff7-450b-abcc-8e374ff891ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.597869 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.736127 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-config-data\") pod \"781f9edc-16e2-44e0-9260-1a8a61fc4950\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.736213 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/781f9edc-16e2-44e0-9260-1a8a61fc4950-logs\") pod \"781f9edc-16e2-44e0-9260-1a8a61fc4950\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.736254 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-combined-ca-bundle\") pod \"781f9edc-16e2-44e0-9260-1a8a61fc4950\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.736456 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgbft\" (UniqueName: \"kubernetes.io/projected/781f9edc-16e2-44e0-9260-1a8a61fc4950-kube-api-access-lgbft\") pod \"781f9edc-16e2-44e0-9260-1a8a61fc4950\" (UID: \"781f9edc-16e2-44e0-9260-1a8a61fc4950\") " Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.736667 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/781f9edc-16e2-44e0-9260-1a8a61fc4950-logs" (OuterVolumeSpecName: "logs") pod "781f9edc-16e2-44e0-9260-1a8a61fc4950" (UID: "781f9edc-16e2-44e0-9260-1a8a61fc4950"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.737072 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/781f9edc-16e2-44e0-9260-1a8a61fc4950-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.758141 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/781f9edc-16e2-44e0-9260-1a8a61fc4950-kube-api-access-lgbft" (OuterVolumeSpecName: "kube-api-access-lgbft") pod "781f9edc-16e2-44e0-9260-1a8a61fc4950" (UID: "781f9edc-16e2-44e0-9260-1a8a61fc4950"). InnerVolumeSpecName "kube-api-access-lgbft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.773450 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "781f9edc-16e2-44e0-9260-1a8a61fc4950" (UID: "781f9edc-16e2-44e0-9260-1a8a61fc4950"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.795150 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-config-data" (OuterVolumeSpecName: "config-data") pod "781f9edc-16e2-44e0-9260-1a8a61fc4950" (UID: "781f9edc-16e2-44e0-9260-1a8a61fc4950"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.838114 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.838152 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/781f9edc-16e2-44e0-9260-1a8a61fc4950-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.838163 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgbft\" (UniqueName: \"kubernetes.io/projected/781f9edc-16e2-44e0-9260-1a8a61fc4950-kube-api-access-lgbft\") on node \"crc\" DevicePath \"\"" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.932619 4843 generic.go:334] "Generic (PLEG): container finished" podID="781f9edc-16e2-44e0-9260-1a8a61fc4950" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" exitCode=0 Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.932679 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.932696 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"781f9edc-16e2-44e0-9260-1a8a61fc4950","Type":"ContainerDied","Data":"91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785"} Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.932739 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"781f9edc-16e2-44e0-9260-1a8a61fc4950","Type":"ContainerDied","Data":"0d59a920f6810e331c38b3e9359152519349751fe6ed6d8e3c05116030126b88"} Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.932759 4843 scope.go:117] "RemoveContainer" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.934077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"8c74421b-0ff7-450b-abcc-8e374ff891ac","Type":"ContainerDied","Data":"7de2f648bb91a55acdcd842126af4a4779d6e22a129eefb89fe20979d2b405a2"} Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.934212 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.934224 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.954781 4843 scope.go:117] "RemoveContainer" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" Mar 14 09:49:58 crc kubenswrapper[4843]: E0314 09:49:58.955581 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785\": container with ID starting with 91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785 not found: ID does not exist" containerID="91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.955638 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785"} err="failed to get container status \"91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785\": rpc error: code = NotFound desc = could not find container \"91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785\": container with ID starting with 91e7bf2d62bc94a86e625ba1e56d5cfcef5be89f683b5177c7f60517b7a70785 not found: ID does not exist" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.955668 4843 scope.go:117] "RemoveContainer" containerID="fc89194c3dae86995ed493f4a9bcad9fe005794878d0df571dbc3e997178bcb8" Mar 14 09:49:58 crc kubenswrapper[4843]: I0314 09:49:58.997193 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.008779 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.018598 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.027605 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.034153 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: E0314 09:49:59.034764 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" containerName="memcached" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.034782 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" containerName="memcached" Mar 14 09:49:59 crc kubenswrapper[4843]: E0314 09:49:59.034829 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-api" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.034836 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-api" Mar 14 09:49:59 crc kubenswrapper[4843]: E0314 09:49:59.034851 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-kuttl-api-log" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.034857 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-kuttl-api-log" Mar 14 09:49:59 crc kubenswrapper[4843]: E0314 09:49:59.034877 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="781f9edc-16e2-44e0-9260-1a8a61fc4950" containerName="watcher-applier" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.034884 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="781f9edc-16e2-44e0-9260-1a8a61fc4950" containerName="watcher-applier" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.035030 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-api" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.035042 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="781f9edc-16e2-44e0-9260-1a8a61fc4950" containerName="watcher-applier" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.035051 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" containerName="memcached" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.035060 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae4d592-7c50-4199-b933-40099a89832d" containerName="watcher-kuttl-api-log" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.047514 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.047611 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.051581 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-memcached-svc" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.051694 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"memcached-memcached-dockercfg-85dpd" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.051757 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"watcher-kuttl-default"/"memcached-config-data" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.061388 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.067437 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.080516 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.081833 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.085434 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.107650 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.108999 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.112768 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.112983 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-internal-svc" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.113130 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-watcher-public-svc" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.130257 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.147836 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.147886 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/08eba4dc-5f7e-47fe-994d-72af55c68a77-memcached-tls-certs\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.147931 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08eba4dc-5f7e-47fe-994d-72af55c68a77-combined-ca-bundle\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.147978 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4wk7\" (UniqueName: \"kubernetes.io/projected/08eba4dc-5f7e-47fe-994d-72af55c68a77-kube-api-access-n4wk7\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.148004 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/08eba4dc-5f7e-47fe-994d-72af55c68a77-kolla-config\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.148057 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.148080 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08eba4dc-5f7e-47fe-994d-72af55c68a77-config-data\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.148106 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.148134 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.148157 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swkf5\" (UniqueName: \"kubernetes.io/projected/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-kube-api-access-swkf5\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.165697 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250014 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250129 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08eba4dc-5f7e-47fe-994d-72af55c68a77-combined-ca-bundle\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250194 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4wk7\" (UniqueName: \"kubernetes.io/projected/08eba4dc-5f7e-47fe-994d-72af55c68a77-kube-api-access-n4wk7\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250221 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250260 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/08eba4dc-5f7e-47fe-994d-72af55c68a77-kolla-config\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250318 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grjwr\" (UniqueName: \"kubernetes.io/projected/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-kube-api-access-grjwr\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250353 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-logs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250398 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250424 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250454 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08eba4dc-5f7e-47fe-994d-72af55c68a77-config-data\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250487 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250510 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250539 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250566 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250598 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250626 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swkf5\" (UniqueName: \"kubernetes.io/projected/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-kube-api-access-swkf5\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250665 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.250725 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/08eba4dc-5f7e-47fe-994d-72af55c68a77-memcached-tls-certs\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.252351 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/08eba4dc-5f7e-47fe-994d-72af55c68a77-kolla-config\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.252357 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08eba4dc-5f7e-47fe-994d-72af55c68a77-config-data\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.253809 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.255847 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/08eba4dc-5f7e-47fe-994d-72af55c68a77-memcached-tls-certs\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.255906 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.255995 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.256423 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.258938 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08eba4dc-5f7e-47fe-994d-72af55c68a77-combined-ca-bundle\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.267638 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4wk7\" (UniqueName: \"kubernetes.io/projected/08eba4dc-5f7e-47fe-994d-72af55c68a77-kube-api-access-n4wk7\") pod \"memcached-0\" (UID: \"08eba4dc-5f7e-47fe-994d-72af55c68a77\") " pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.267844 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swkf5\" (UniqueName: \"kubernetes.io/projected/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-kube-api-access-swkf5\") pod \"watcher-kuttl-applier-0\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.349016 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae4d592-7c50-4199-b933-40099a89832d" path="/var/lib/kubelet/pods/2ae4d592-7c50-4199-b933-40099a89832d/volumes" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.349652 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="781f9edc-16e2-44e0-9260-1a8a61fc4950" path="/var/lib/kubelet/pods/781f9edc-16e2-44e0-9260-1a8a61fc4950/volumes" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.350141 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c74421b-0ff7-450b-abcc-8e374ff891ac" path="/var/lib/kubelet/pods/8c74421b-0ff7-450b-abcc-8e374ff891ac/volumes" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.351809 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.351869 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.351909 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.351938 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grjwr\" (UniqueName: \"kubernetes.io/projected/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-kube-api-access-grjwr\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.351961 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-logs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.351988 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.352018 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.352038 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.352537 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-logs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.355478 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.355486 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.355863 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.356118 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.357078 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-public-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.357590 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-internal-tls-certs\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.371818 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grjwr\" (UniqueName: \"kubernetes.io/projected/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-kube-api-access-grjwr\") pod \"watcher-kuttl-api-0\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.404586 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/memcached-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.423101 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.433998 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.904872 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: W0314 09:49:59.906219 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda31aec38_dc3f_4143_8b4d_bc7b61f169ea.slice/crio-dfa87d670635984792f5471ad13f3d93cc15e7cf3e91f6b25332510e57315dd9 WatchSource:0}: Error finding container dfa87d670635984792f5471ad13f3d93cc15e7cf3e91f6b25332510e57315dd9: Status 404 returned error can't find the container with id dfa87d670635984792f5471ad13f3d93cc15e7cf3e91f6b25332510e57315dd9 Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.920439 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.957423 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"a31aec38-dc3f-4143-8b4d-bc7b61f169ea","Type":"ContainerStarted","Data":"dfa87d670635984792f5471ad13f3d93cc15e7cf3e91f6b25332510e57315dd9"} Mar 14 09:49:59 crc kubenswrapper[4843]: I0314 09:49:59.959082 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f","Type":"ContainerStarted","Data":"bc6f66baa7f4777966b7bdfc36a3e7f734e6e4f8609062ee337a80e623efa6fd"} Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.028051 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/memcached-0"] Mar 14 09:50:00 crc kubenswrapper[4843]: W0314 09:50:00.033704 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08eba4dc_5f7e_47fe_994d_72af55c68a77.slice/crio-c567a36801fec96b0b21133da068caefddcac88f09debe5dd252bf1e05310cb5 WatchSource:0}: Error finding container c567a36801fec96b0b21133da068caefddcac88f09debe5dd252bf1e05310cb5: Status 404 returned error can't find the container with id c567a36801fec96b0b21133da068caefddcac88f09debe5dd252bf1e05310cb5 Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.142934 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558030-x5ltd"] Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.144844 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.148621 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.148761 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.148726 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.154701 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558030-x5ltd"] Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.271985 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgwg7\" (UniqueName: \"kubernetes.io/projected/f61b3db5-0fa4-4809-b8b3-8974f6363775-kube-api-access-dgwg7\") pod \"auto-csr-approver-29558030-x5ltd\" (UID: \"f61b3db5-0fa4-4809-b8b3-8974f6363775\") " pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.373349 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgwg7\" (UniqueName: \"kubernetes.io/projected/f61b3db5-0fa4-4809-b8b3-8974f6363775-kube-api-access-dgwg7\") pod \"auto-csr-approver-29558030-x5ltd\" (UID: \"f61b3db5-0fa4-4809-b8b3-8974f6363775\") " pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.391147 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgwg7\" (UniqueName: \"kubernetes.io/projected/f61b3db5-0fa4-4809-b8b3-8974f6363775-kube-api-access-dgwg7\") pod \"auto-csr-approver-29558030-x5ltd\" (UID: \"f61b3db5-0fa4-4809-b8b3-8974f6363775\") " pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.479118 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.974628 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"08eba4dc-5f7e-47fe-994d-72af55c68a77","Type":"ContainerStarted","Data":"03581c0a39a905bc2e897a0676c8906a83a7daed9ee15aaa90d631f32d91f2e1"} Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.975004 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/memcached-0" event={"ID":"08eba4dc-5f7e-47fe-994d-72af55c68a77","Type":"ContainerStarted","Data":"c567a36801fec96b0b21133da068caefddcac88f09debe5dd252bf1e05310cb5"} Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.975646 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/memcached-0" Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.984542 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"a31aec38-dc3f-4143-8b4d-bc7b61f169ea","Type":"ContainerStarted","Data":"5626a70f5fca2153346a27da538352451986038107b6c35b5f5007104f430a8a"} Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.989027 4843 generic.go:334] "Generic (PLEG): container finished" podID="ffab3465-469a-45ae-8633-0d22589ae869" containerID="39518908e826d8a0f2bf3e7f8ca5097f2aa3bddfc661fced13a318202bf7f790" exitCode=0 Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.989132 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" event={"ID":"ffab3465-469a-45ae-8633-0d22589ae869","Type":"ContainerDied","Data":"39518908e826d8a0f2bf3e7f8ca5097f2aa3bddfc661fced13a318202bf7f790"} Mar 14 09:50:00 crc kubenswrapper[4843]: I0314 09:50:00.995242 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/memcached-0" podStartSLOduration=2.995215194 podStartE2EDuration="2.995215194s" podCreationTimestamp="2026-03-14 09:49:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:50:00.993418472 +0000 UTC m=+2368.306029600" watchObservedRunningTime="2026-03-14 09:50:00.995215194 +0000 UTC m=+2368.307826322" Mar 14 09:50:01 crc kubenswrapper[4843]: I0314 09:50:01.016698 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f","Type":"ContainerStarted","Data":"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0"} Mar 14 09:50:01 crc kubenswrapper[4843]: I0314 09:50:01.017147 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f","Type":"ContainerStarted","Data":"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292"} Mar 14 09:50:01 crc kubenswrapper[4843]: I0314 09:50:01.017672 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:01 crc kubenswrapper[4843]: I0314 09:50:01.041792 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=3.041765404 podStartE2EDuration="3.041765404s" podCreationTimestamp="2026-03-14 09:49:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:50:01.038033445 +0000 UTC m=+2368.350644653" watchObservedRunningTime="2026-03-14 09:50:01.041765404 +0000 UTC m=+2368.354376532" Mar 14 09:50:01 crc kubenswrapper[4843]: I0314 09:50:01.061024 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.060996572 podStartE2EDuration="2.060996572s" podCreationTimestamp="2026-03-14 09:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:50:01.056021454 +0000 UTC m=+2368.368632582" watchObservedRunningTime="2026-03-14 09:50:01.060996572 +0000 UTC m=+2368.373607700" Mar 14 09:50:01 crc kubenswrapper[4843]: I0314 09:50:01.099521 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558030-x5ltd"] Mar 14 09:50:01 crc kubenswrapper[4843]: E0314 09:50:01.357098 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Mar 14 09:50:01 crc kubenswrapper[4843]: E0314 09:50:01.358383 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Mar 14 09:50:01 crc kubenswrapper[4843]: E0314 09:50:01.359185 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Mar 14 09:50:01 crc kubenswrapper[4843]: E0314 09:50:01.359231 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="96005028-fce4-421d-8a81-85bf81e1742a" containerName="watcher-decision-engine" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.029907 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" event={"ID":"f61b3db5-0fa4-4809-b8b3-8974f6363775","Type":"ContainerStarted","Data":"9b0e8a216ab94f1c944898300f887e7c3074fe998b069baca1ea532a70324cf4"} Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.339155 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:50:02 crc kubenswrapper[4843]: E0314 09:50:02.339859 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.418187 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.520870 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-fernet-keys\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.520941 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-credential-keys\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.520995 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-cert-memcached-mtls\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.521119 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-scripts\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.521181 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4zfm\" (UniqueName: \"kubernetes.io/projected/ffab3465-469a-45ae-8633-0d22589ae869-kube-api-access-p4zfm\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.521249 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-combined-ca-bundle\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.521321 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-config-data\") pod \"ffab3465-469a-45ae-8633-0d22589ae869\" (UID: \"ffab3465-469a-45ae-8633-0d22589ae869\") " Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.527567 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.527653 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-scripts" (OuterVolumeSpecName: "scripts") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.527796 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.527976 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffab3465-469a-45ae-8633-0d22589ae869-kube-api-access-p4zfm" (OuterVolumeSpecName: "kube-api-access-p4zfm") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "kube-api-access-p4zfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.552364 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-config-data" (OuterVolumeSpecName: "config-data") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.554209 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.584326 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "ffab3465-469a-45ae-8633-0d22589ae869" (UID: "ffab3465-469a-45ae-8633-0d22589ae869"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624199 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624238 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4zfm\" (UniqueName: \"kubernetes.io/projected/ffab3465-469a-45ae-8633-0d22589ae869-kube-api-access-p4zfm\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624250 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624261 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624286 4843 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624294 4843 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:02 crc kubenswrapper[4843]: I0314 09:50:02.624303 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/ffab3465-469a-45ae-8633-0d22589ae869-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.052529 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" event={"ID":"ffab3465-469a-45ae-8633-0d22589ae869","Type":"ContainerDied","Data":"b21057d64bbadf46a77b0eb3a7e7f98002d06a9e310ad4763f82b69433f346bd"} Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.052633 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b21057d64bbadf46a77b0eb3a7e7f98002d06a9e310ad4763f82b69433f346bd" Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.052925 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-bootstrap-gw9wj" Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.053966 4843 generic.go:334] "Generic (PLEG): container finished" podID="f61b3db5-0fa4-4809-b8b3-8974f6363775" containerID="08441fe3385a19c95b327c82bd4507f5309a6a031a84ec3e29c38a1ecd019cd5" exitCode=0 Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.054020 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" event={"ID":"f61b3db5-0fa4-4809-b8b3-8974f6363775","Type":"ContainerDied","Data":"08441fe3385a19c95b327c82bd4507f5309a6a031a84ec3e29c38a1ecd019cd5"} Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.054061 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:50:03 crc kubenswrapper[4843]: I0314 09:50:03.406221 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.064902 4843 generic.go:334] "Generic (PLEG): container finished" podID="96005028-fce4-421d-8a81-85bf81e1742a" containerID="59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c" exitCode=0 Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.064994 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"96005028-fce4-421d-8a81-85bf81e1742a","Type":"ContainerDied","Data":"59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c"} Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.425061 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.434187 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.531776 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.659080 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.664338 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-custom-prometheus-ca\") pod \"96005028-fce4-421d-8a81-85bf81e1742a\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.664498 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96005028-fce4-421d-8a81-85bf81e1742a-logs\") pod \"96005028-fce4-421d-8a81-85bf81e1742a\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.664538 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-config-data\") pod \"96005028-fce4-421d-8a81-85bf81e1742a\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.664630 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-combined-ca-bundle\") pod \"96005028-fce4-421d-8a81-85bf81e1742a\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.664662 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9mnz\" (UniqueName: \"kubernetes.io/projected/96005028-fce4-421d-8a81-85bf81e1742a-kube-api-access-z9mnz\") pod \"96005028-fce4-421d-8a81-85bf81e1742a\" (UID: \"96005028-fce4-421d-8a81-85bf81e1742a\") " Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.666090 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96005028-fce4-421d-8a81-85bf81e1742a-logs" (OuterVolumeSpecName: "logs") pod "96005028-fce4-421d-8a81-85bf81e1742a" (UID: "96005028-fce4-421d-8a81-85bf81e1742a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.692736 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96005028-fce4-421d-8a81-85bf81e1742a" (UID: "96005028-fce4-421d-8a81-85bf81e1742a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.717776 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96005028-fce4-421d-8a81-85bf81e1742a-kube-api-access-z9mnz" (OuterVolumeSpecName: "kube-api-access-z9mnz") pod "96005028-fce4-421d-8a81-85bf81e1742a" (UID: "96005028-fce4-421d-8a81-85bf81e1742a"). InnerVolumeSpecName "kube-api-access-z9mnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.766294 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgwg7\" (UniqueName: \"kubernetes.io/projected/f61b3db5-0fa4-4809-b8b3-8974f6363775-kube-api-access-dgwg7\") pod \"f61b3db5-0fa4-4809-b8b3-8974f6363775\" (UID: \"f61b3db5-0fa4-4809-b8b3-8974f6363775\") " Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.766741 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96005028-fce4-421d-8a81-85bf81e1742a-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.766756 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.766769 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9mnz\" (UniqueName: \"kubernetes.io/projected/96005028-fce4-421d-8a81-85bf81e1742a-kube-api-access-z9mnz\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.770501 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "96005028-fce4-421d-8a81-85bf81e1742a" (UID: "96005028-fce4-421d-8a81-85bf81e1742a"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.772561 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f61b3db5-0fa4-4809-b8b3-8974f6363775-kube-api-access-dgwg7" (OuterVolumeSpecName: "kube-api-access-dgwg7") pod "f61b3db5-0fa4-4809-b8b3-8974f6363775" (UID: "f61b3db5-0fa4-4809-b8b3-8974f6363775"). InnerVolumeSpecName "kube-api-access-dgwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.801499 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-config-data" (OuterVolumeSpecName: "config-data") pod "96005028-fce4-421d-8a81-85bf81e1742a" (UID: "96005028-fce4-421d-8a81-85bf81e1742a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.874705 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.874760 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgwg7\" (UniqueName: \"kubernetes.io/projected/f61b3db5-0fa4-4809-b8b3-8974f6363775-kube-api-access-dgwg7\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:04 crc kubenswrapper[4843]: I0314 09:50:04.874770 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/96005028-fce4-421d-8a81-85bf81e1742a-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.076203 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"96005028-fce4-421d-8a81-85bf81e1742a","Type":"ContainerDied","Data":"b97ea8c5d8e99b54b7cdb7cc65e39eac66526f790fb1c72122be7bfe351aa076"} Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.076559 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.077547 4843 scope.go:117] "RemoveContainer" containerID="59c5485aa78369a42ca50a2eb5d258991ece73c04d9ac7db51fe478485251f9c" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.080466 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.080515 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558030-x5ltd" event={"ID":"f61b3db5-0fa4-4809-b8b3-8974f6363775","Type":"ContainerDied","Data":"9b0e8a216ab94f1c944898300f887e7c3074fe998b069baca1ea532a70324cf4"} Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.080550 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b0e8a216ab94f1c944898300f887e7c3074fe998b069baca1ea532a70324cf4" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.147001 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.171067 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186101 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:50:05 crc kubenswrapper[4843]: E0314 09:50:05.186659 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f61b3db5-0fa4-4809-b8b3-8974f6363775" containerName="oc" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186677 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f61b3db5-0fa4-4809-b8b3-8974f6363775" containerName="oc" Mar 14 09:50:05 crc kubenswrapper[4843]: E0314 09:50:05.186700 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffab3465-469a-45ae-8633-0d22589ae869" containerName="keystone-bootstrap" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186709 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffab3465-469a-45ae-8633-0d22589ae869" containerName="keystone-bootstrap" Mar 14 09:50:05 crc kubenswrapper[4843]: E0314 09:50:05.186733 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96005028-fce4-421d-8a81-85bf81e1742a" containerName="watcher-decision-engine" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186741 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="96005028-fce4-421d-8a81-85bf81e1742a" containerName="watcher-decision-engine" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186940 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="96005028-fce4-421d-8a81-85bf81e1742a" containerName="watcher-decision-engine" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186953 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffab3465-469a-45ae-8633-0d22589ae869" containerName="keystone-bootstrap" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.186967 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f61b3db5-0fa4-4809-b8b3-8974f6363775" containerName="oc" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.187830 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.191281 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.199114 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.288941 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.289039 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.289088 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f31484f2-c78d-4896-ab8d-c6aa331c5255-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.289122 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.289226 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8brn\" (UniqueName: \"kubernetes.io/projected/f31484f2-c78d-4896-ab8d-c6aa331c5255-kube-api-access-t8brn\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.289250 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.350112 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96005028-fce4-421d-8a81-85bf81e1742a" path="/var/lib/kubelet/pods/96005028-fce4-421d-8a81-85bf81e1742a/volumes" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.391329 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.391431 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.391483 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f31484f2-c78d-4896-ab8d-c6aa331c5255-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.391536 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.391599 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8brn\" (UniqueName: \"kubernetes.io/projected/f31484f2-c78d-4896-ab8d-c6aa331c5255-kube-api-access-t8brn\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.391629 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.392629 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f31484f2-c78d-4896-ab8d-c6aa331c5255-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.396298 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.396427 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.397869 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.397876 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.433857 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8brn\" (UniqueName: \"kubernetes.io/projected/f31484f2-c78d-4896-ab8d-c6aa331c5255-kube-api-access-t8brn\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.506247 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.756757 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558024-mgk5q"] Mar 14 09:50:05 crc kubenswrapper[4843]: I0314 09:50:05.766118 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558024-mgk5q"] Mar 14 09:50:06 crc kubenswrapper[4843]: I0314 09:50:06.030125 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:50:06 crc kubenswrapper[4843]: W0314 09:50:06.033731 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf31484f2_c78d_4896_ab8d_c6aa331c5255.slice/crio-cb843b526d8c52e8e60f6f72c0219ae7f394e00e2706aa4ed78adf2daca0f095 WatchSource:0}: Error finding container cb843b526d8c52e8e60f6f72c0219ae7f394e00e2706aa4ed78adf2daca0f095: Status 404 returned error can't find the container with id cb843b526d8c52e8e60f6f72c0219ae7f394e00e2706aa4ed78adf2daca0f095 Mar 14 09:50:06 crc kubenswrapper[4843]: I0314 09:50:06.092549 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"f31484f2-c78d-4896-ab8d-c6aa331c5255","Type":"ContainerStarted","Data":"cb843b526d8c52e8e60f6f72c0219ae7f394e00e2706aa4ed78adf2daca0f095"} Mar 14 09:50:07 crc kubenswrapper[4843]: I0314 09:50:07.101090 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"f31484f2-c78d-4896-ab8d-c6aa331c5255","Type":"ContainerStarted","Data":"5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36"} Mar 14 09:50:07 crc kubenswrapper[4843]: I0314 09:50:07.123496 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.123472434 podStartE2EDuration="2.123472434s" podCreationTimestamp="2026-03-14 09:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:50:07.116884828 +0000 UTC m=+2374.429495986" watchObservedRunningTime="2026-03-14 09:50:07.123472434 +0000 UTC m=+2374.436083602" Mar 14 09:50:07 crc kubenswrapper[4843]: I0314 09:50:07.350882 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36fd9387-d70a-4b8d-9d0a-9f6b507432df" path="/var/lib/kubelet/pods/36fd9387-d70a-4b8d-9d0a-9f6b507432df/volumes" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.406935 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/memcached-0" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.424668 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.434294 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.449636 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.463748 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.539600 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-7f5697c5fb-hqksz"] Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.540552 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.558191 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-7f5697c5fb-hqksz"] Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666065 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-credential-keys\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666151 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-scripts\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666187 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rngw9\" (UniqueName: \"kubernetes.io/projected/494a65cb-9ba0-448d-944c-7040f8e30bf7-kube-api-access-rngw9\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666221 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-public-tls-certs\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666335 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-combined-ca-bundle\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666359 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-fernet-keys\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666413 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-cert-memcached-mtls\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666493 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-internal-tls-certs\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.666539 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-config-data\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768409 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rngw9\" (UniqueName: \"kubernetes.io/projected/494a65cb-9ba0-448d-944c-7040f8e30bf7-kube-api-access-rngw9\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768460 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-public-tls-certs\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768516 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-combined-ca-bundle\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768536 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-fernet-keys\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768570 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-cert-memcached-mtls\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768588 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-internal-tls-certs\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768607 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-config-data\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768636 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-credential-keys\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.768672 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-scripts\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.776751 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-scripts\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.777119 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-credential-keys\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.777791 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-fernet-keys\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.777830 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-config-data\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.779959 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-combined-ca-bundle\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.780936 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-internal-tls-certs\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.782956 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-cert-memcached-mtls\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.793161 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rngw9\" (UniqueName: \"kubernetes.io/projected/494a65cb-9ba0-448d-944c-7040f8e30bf7-kube-api-access-rngw9\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.794267 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/494a65cb-9ba0-448d-944c-7040f8e30bf7-public-tls-certs\") pod \"keystone-7f5697c5fb-hqksz\" (UID: \"494a65cb-9ba0-448d-944c-7040f8e30bf7\") " pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:09 crc kubenswrapper[4843]: I0314 09:50:09.857714 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:10 crc kubenswrapper[4843]: I0314 09:50:10.137334 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:10 crc kubenswrapper[4843]: I0314 09:50:10.164890 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:50:10 crc kubenswrapper[4843]: I0314 09:50:10.291307 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-7f5697c5fb-hqksz"] Mar 14 09:50:10 crc kubenswrapper[4843]: W0314 09:50:10.291439 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod494a65cb_9ba0_448d_944c_7040f8e30bf7.slice/crio-f928f3670476e36022941d8e1fc60d551a867c369406e1064b324b0c2f8dcda8 WatchSource:0}: Error finding container f928f3670476e36022941d8e1fc60d551a867c369406e1064b324b0c2f8dcda8: Status 404 returned error can't find the container with id f928f3670476e36022941d8e1fc60d551a867c369406e1064b324b0c2f8dcda8 Mar 14 09:50:11 crc kubenswrapper[4843]: I0314 09:50:11.134526 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" event={"ID":"494a65cb-9ba0-448d-944c-7040f8e30bf7","Type":"ContainerStarted","Data":"b9e2e8417c1da1bb12ec1a058eb157fab616a3e09137f9d57080945003b52d19"} Mar 14 09:50:11 crc kubenswrapper[4843]: I0314 09:50:11.134991 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" event={"ID":"494a65cb-9ba0-448d-944c-7040f8e30bf7","Type":"ContainerStarted","Data":"f928f3670476e36022941d8e1fc60d551a867c369406e1064b324b0c2f8dcda8"} Mar 14 09:50:11 crc kubenswrapper[4843]: I0314 09:50:11.159235 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" podStartSLOduration=2.159210456 podStartE2EDuration="2.159210456s" podCreationTimestamp="2026-03-14 09:50:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:50:11.153799706 +0000 UTC m=+2378.466410834" watchObservedRunningTime="2026-03-14 09:50:11.159210456 +0000 UTC m=+2378.471821584" Mar 14 09:50:11 crc kubenswrapper[4843]: I0314 09:50:11.648902 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:50:12 crc kubenswrapper[4843]: I0314 09:50:12.142022 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-kuttl-api-log" containerID="cri-o://afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292" gracePeriod=30 Mar 14 09:50:12 crc kubenswrapper[4843]: I0314 09:50:12.142086 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-api" containerID="cri-o://dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0" gracePeriod=30 Mar 14 09:50:12 crc kubenswrapper[4843]: I0314 09:50:12.143239 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:12 crc kubenswrapper[4843]: I0314 09:50:12.962803 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031103 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-internal-tls-certs\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031194 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-public-tls-certs\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031292 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-config-data\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031333 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-custom-prometheus-ca\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031367 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-logs\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031413 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grjwr\" (UniqueName: \"kubernetes.io/projected/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-kube-api-access-grjwr\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031434 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-cert-memcached-mtls\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.031461 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-combined-ca-bundle\") pod \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\" (UID: \"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f\") " Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.032914 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-logs" (OuterVolumeSpecName: "logs") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.039506 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-kube-api-access-grjwr" (OuterVolumeSpecName: "kube-api-access-grjwr") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "kube-api-access-grjwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.070287 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.089475 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.095716 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.095738 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-config-data" (OuterVolumeSpecName: "config-data") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.123570 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.134727 4843 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.134977 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.135052 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.135119 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.135203 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grjwr\" (UniqueName: \"kubernetes.io/projected/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-kube-api-access-grjwr\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.135302 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.135394 4843 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.136493 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" (UID: "39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.150346 4843 generic.go:334] "Generic (PLEG): container finished" podID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerID="dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0" exitCode=0 Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.150375 4843 generic.go:334] "Generic (PLEG): container finished" podID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerID="afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292" exitCode=143 Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.151118 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.151172 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f","Type":"ContainerDied","Data":"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0"} Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.151197 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f","Type":"ContainerDied","Data":"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292"} Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.151209 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f","Type":"ContainerDied","Data":"bc6f66baa7f4777966b7bdfc36a3e7f734e6e4f8609062ee337a80e623efa6fd"} Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.151226 4843 scope.go:117] "RemoveContainer" containerID="dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.173150 4843 scope.go:117] "RemoveContainer" containerID="afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.189911 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.197948 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.202924 4843 scope.go:117] "RemoveContainer" containerID="dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0" Mar 14 09:50:13 crc kubenswrapper[4843]: E0314 09:50:13.203576 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0\": container with ID starting with dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0 not found: ID does not exist" containerID="dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.203618 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0"} err="failed to get container status \"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0\": rpc error: code = NotFound desc = could not find container \"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0\": container with ID starting with dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0 not found: ID does not exist" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.203648 4843 scope.go:117] "RemoveContainer" containerID="afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292" Mar 14 09:50:13 crc kubenswrapper[4843]: E0314 09:50:13.205301 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292\": container with ID starting with afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292 not found: ID does not exist" containerID="afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.205343 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292"} err="failed to get container status \"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292\": rpc error: code = NotFound desc = could not find container \"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292\": container with ID starting with afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292 not found: ID does not exist" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.205371 4843 scope.go:117] "RemoveContainer" containerID="dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.205612 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0"} err="failed to get container status \"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0\": rpc error: code = NotFound desc = could not find container \"dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0\": container with ID starting with dc3c841fb2dfe983eb29452b7fa0ec0475fcb4b70599bc8830ea43c9017f2bd0 not found: ID does not exist" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.205638 4843 scope.go:117] "RemoveContainer" containerID="afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.206065 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:50:13 crc kubenswrapper[4843]: E0314 09:50:13.206509 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-api" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.206594 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-api" Mar 14 09:50:13 crc kubenswrapper[4843]: E0314 09:50:13.206689 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-kuttl-api-log" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.206744 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-kuttl-api-log" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.206125 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292"} err="failed to get container status \"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292\": rpc error: code = NotFound desc = could not find container \"afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292\": container with ID starting with afbf0200cd1c4888adbdd15eab85e895b4757bacf8673574871849f00d91d292 not found: ID does not exist" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.207181 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-api" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.207287 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" containerName="watcher-kuttl-api-log" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.208256 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.216176 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.229709 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.236817 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.337922 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-logs\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.337971 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7lvm\" (UniqueName: \"kubernetes.io/projected/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-kube-api-access-t7lvm\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.338050 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.338106 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.338152 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.338261 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.368455 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f" path="/var/lib/kubelet/pods/39bb167a-e184-4dd4-8d8a-cd6ae8bf1a6f/volumes" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.439637 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-logs\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.439735 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7lvm\" (UniqueName: \"kubernetes.io/projected/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-kube-api-access-t7lvm\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.440121 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.440178 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.440211 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.440311 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.441508 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-logs\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.450094 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.451098 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.456815 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.469016 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.471764 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7lvm\" (UniqueName: \"kubernetes.io/projected/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-kube-api-access-t7lvm\") pod \"watcher-kuttl-api-0\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.528427 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:13 crc kubenswrapper[4843]: I0314 09:50:13.964145 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:50:13 crc kubenswrapper[4843]: W0314 09:50:13.975366 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58112b9d_0e99_4e9b_9b6a_655c1ba03bf7.slice/crio-0128996aa9a8f9be820eb9a3d2e8cbc5ad05be152f0eba804ae52c8b79b24c36 WatchSource:0}: Error finding container 0128996aa9a8f9be820eb9a3d2e8cbc5ad05be152f0eba804ae52c8b79b24c36: Status 404 returned error can't find the container with id 0128996aa9a8f9be820eb9a3d2e8cbc5ad05be152f0eba804ae52c8b79b24c36 Mar 14 09:50:14 crc kubenswrapper[4843]: I0314 09:50:14.166019 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7","Type":"ContainerStarted","Data":"e7c063ed07886f5e11ea86ca7dcb680b4f6c864b7fd9965fec71ba1e890a85e4"} Mar 14 09:50:14 crc kubenswrapper[4843]: I0314 09:50:14.166395 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7","Type":"ContainerStarted","Data":"0128996aa9a8f9be820eb9a3d2e8cbc5ad05be152f0eba804ae52c8b79b24c36"} Mar 14 09:50:15 crc kubenswrapper[4843]: I0314 09:50:15.176951 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7","Type":"ContainerStarted","Data":"0a82902b88a501520e4b9cadbb9703a08ea48cf796e2d422fbb8ea815958b146"} Mar 14 09:50:15 crc kubenswrapper[4843]: I0314 09:50:15.177617 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:15 crc kubenswrapper[4843]: I0314 09:50:15.202851 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.202825694 podStartE2EDuration="2.202825694s" podCreationTimestamp="2026-03-14 09:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:50:15.198912571 +0000 UTC m=+2382.511523709" watchObservedRunningTime="2026-03-14 09:50:15.202825694 +0000 UTC m=+2382.515436832" Mar 14 09:50:15 crc kubenswrapper[4843]: I0314 09:50:15.507314 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:15 crc kubenswrapper[4843]: I0314 09:50:15.536432 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:16 crc kubenswrapper[4843]: I0314 09:50:16.185413 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:16 crc kubenswrapper[4843]: I0314 09:50:16.221407 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:50:16 crc kubenswrapper[4843]: I0314 09:50:16.339317 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:50:16 crc kubenswrapper[4843]: E0314 09:50:16.339506 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:50:16 crc kubenswrapper[4843]: I0314 09:50:16.487030 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:17 crc kubenswrapper[4843]: I0314 09:50:17.591578 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:18 crc kubenswrapper[4843]: I0314 09:50:18.528955 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:23 crc kubenswrapper[4843]: I0314 09:50:23.529171 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:23 crc kubenswrapper[4843]: I0314 09:50:23.536982 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:24 crc kubenswrapper[4843]: I0314 09:50:24.250731 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:50:29 crc kubenswrapper[4843]: I0314 09:50:29.339019 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:50:29 crc kubenswrapper[4843]: E0314 09:50:29.339862 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:50:41 crc kubenswrapper[4843]: I0314 09:50:41.341450 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:50:41 crc kubenswrapper[4843]: E0314 09:50:41.342216 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:50:41 crc kubenswrapper[4843]: I0314 09:50:41.511139 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/keystone-7f5697c5fb-hqksz" Mar 14 09:50:41 crc kubenswrapper[4843]: I0314 09:50:41.570550 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-577555cd46-j7w69"] Mar 14 09:50:41 crc kubenswrapper[4843]: I0314 09:50:41.570883 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" podUID="cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" containerName="keystone-api" containerID="cri-o://80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8" gracePeriod=30 Mar 14 09:50:44 crc kubenswrapper[4843]: E0314 09:50:44.904384 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb9f076c_8aa4_41c0_aa84_6a1eb0239a86.slice/crio-80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8.scope\": RecentStats: unable to find data in memory cache]" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.323807 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.413857 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-config-data\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.413933 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vpcl\" (UniqueName: \"kubernetes.io/projected/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-kube-api-access-6vpcl\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.414063 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-internal-tls-certs\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.414083 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-scripts\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.414104 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-combined-ca-bundle\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.414142 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-credential-keys\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.414186 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-public-tls-certs\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.414213 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-fernet-keys\") pod \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\" (UID: \"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86\") " Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.419579 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-kube-api-access-6vpcl" (OuterVolumeSpecName: "kube-api-access-6vpcl") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "kube-api-access-6vpcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.420345 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.425282 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-scripts" (OuterVolumeSpecName: "scripts") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.426718 4843 generic.go:334] "Generic (PLEG): container finished" podID="cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" containerID="80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8" exitCode=0 Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.426763 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" event={"ID":"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86","Type":"ContainerDied","Data":"80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8"} Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.426793 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" event={"ID":"cb9f076c-8aa4-41c0-aa84-6a1eb0239a86","Type":"ContainerDied","Data":"4b8e10f76dc986698b74dcad7d5529c7a7cb5d40edcfc206354ac03cc45de422"} Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.426812 4843 scope.go:117] "RemoveContainer" containerID="80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.426946 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-577555cd46-j7w69" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.441409 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.443447 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-config-data" (OuterVolumeSpecName: "config-data") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.455881 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.457918 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.491023 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" (UID: "cb9f076c-8aa4-41c0-aa84-6a1eb0239a86"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515717 4843 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515752 4843 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515763 4843 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515771 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515780 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vpcl\" (UniqueName: \"kubernetes.io/projected/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-kube-api-access-6vpcl\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515791 4843 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515799 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.515808 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.518595 4843 scope.go:117] "RemoveContainer" containerID="80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8" Mar 14 09:50:45 crc kubenswrapper[4843]: E0314 09:50:45.519110 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8\": container with ID starting with 80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8 not found: ID does not exist" containerID="80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.519163 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8"} err="failed to get container status \"80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8\": rpc error: code = NotFound desc = could not find container \"80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8\": container with ID starting with 80eba9c9eb3781125c86cceef1d5e029a343b0ca437a55abb613bb02724837b8 not found: ID does not exist" Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.572370 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.572784 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-central-agent" containerID="cri-o://0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa" gracePeriod=30 Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.572847 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="proxy-httpd" containerID="cri-o://13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75" gracePeriod=30 Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.572844 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="sg-core" containerID="cri-o://d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f" gracePeriod=30 Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.572896 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-notification-agent" containerID="cri-o://f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7" gracePeriod=30 Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.753462 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-577555cd46-j7w69"] Mar 14 09:50:45 crc kubenswrapper[4843]: I0314 09:50:45.774840 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-577555cd46-j7w69"] Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.438242 4843 generic.go:334] "Generic (PLEG): container finished" podID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerID="13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75" exitCode=0 Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.438647 4843 generic.go:334] "Generic (PLEG): container finished" podID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerID="d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f" exitCode=2 Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.438665 4843 generic.go:334] "Generic (PLEG): container finished" podID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerID="0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa" exitCode=0 Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.438498 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerDied","Data":"13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75"} Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.438760 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerDied","Data":"d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f"} Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.438776 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerDied","Data":"0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa"} Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.475484 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": dial tcp 10.217.0.196:3000: connect: connection refused" Mar 14 09:50:46 crc kubenswrapper[4843]: I0314 09:50:46.866380 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041360 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-scripts\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041419 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-log-httpd\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041474 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-combined-ca-bundle\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041536 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-run-httpd\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041554 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfvhm\" (UniqueName: \"kubernetes.io/projected/28ca6ab9-a606-4631-8714-dbca98f6cfcd-kube-api-access-gfvhm\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041579 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-ceilometer-tls-certs\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041630 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-config-data\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.041650 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-sg-core-conf-yaml\") pod \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\" (UID: \"28ca6ab9-a606-4631-8714-dbca98f6cfcd\") " Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.043047 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.043113 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.046939 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-scripts" (OuterVolumeSpecName: "scripts") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.047951 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28ca6ab9-a606-4631-8714-dbca98f6cfcd-kube-api-access-gfvhm" (OuterVolumeSpecName: "kube-api-access-gfvhm") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "kube-api-access-gfvhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.066978 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.105347 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.108841 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.132358 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-config-data" (OuterVolumeSpecName: "config-data") pod "28ca6ab9-a606-4631-8714-dbca98f6cfcd" (UID: "28ca6ab9-a606-4631-8714-dbca98f6cfcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.162967 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.162999 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.163009 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.163020 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/28ca6ab9-a606-4631-8714-dbca98f6cfcd-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.163028 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfvhm\" (UniqueName: \"kubernetes.io/projected/28ca6ab9-a606-4631-8714-dbca98f6cfcd-kube-api-access-gfvhm\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.163036 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.163045 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.163053 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/28ca6ab9-a606-4631-8714-dbca98f6cfcd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.355851 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" path="/var/lib/kubelet/pods/cb9f076c-8aa4-41c0-aa84-6a1eb0239a86/volumes" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.450737 4843 generic.go:334] "Generic (PLEG): container finished" podID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerID="f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7" exitCode=0 Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.450776 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerDied","Data":"f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7"} Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.450818 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"28ca6ab9-a606-4631-8714-dbca98f6cfcd","Type":"ContainerDied","Data":"81a9075c3f68009b6ac980491dbaddafcacdd4976e61a58c22a4d9a6c6aa4fa6"} Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.450836 4843 scope.go:117] "RemoveContainer" containerID="13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.450847 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.479830 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.488163 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.489955 4843 scope.go:117] "RemoveContainer" containerID="d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.514038 4843 scope.go:117] "RemoveContainer" containerID="f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.515414 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.515725 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" containerName="keystone-api" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516397 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" containerName="keystone-api" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.516467 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="proxy-httpd" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516481 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="proxy-httpd" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.516500 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="sg-core" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516512 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="sg-core" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.516556 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-notification-agent" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516575 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-notification-agent" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.516598 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-central-agent" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516607 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-central-agent" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516936 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-notification-agent" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516957 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="proxy-httpd" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516972 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9f076c-8aa4-41c0-aa84-6a1eb0239a86" containerName="keystone-api" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.516985 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="sg-core" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.517029 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" containerName="ceilometer-central-agent" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.519684 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.526394 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.527151 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.528631 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.531755 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.535295 4843 scope.go:117] "RemoveContainer" containerID="0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.560261 4843 scope.go:117] "RemoveContainer" containerID="13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.560905 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75\": container with ID starting with 13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75 not found: ID does not exist" containerID="13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.560947 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75"} err="failed to get container status \"13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75\": rpc error: code = NotFound desc = could not find container \"13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75\": container with ID starting with 13c2a41cab12f399cf4303fce8368a948b7260ab01da209c2e5efc1a2a943f75 not found: ID does not exist" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.560974 4843 scope.go:117] "RemoveContainer" containerID="d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.561460 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f\": container with ID starting with d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f not found: ID does not exist" containerID="d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.561483 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f"} err="failed to get container status \"d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f\": rpc error: code = NotFound desc = could not find container \"d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f\": container with ID starting with d91c39715fe287b6b235193279e13cde5b8f429b6d688e22e6b7c0b8733cbe6f not found: ID does not exist" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.561495 4843 scope.go:117] "RemoveContainer" containerID="f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.561696 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7\": container with ID starting with f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7 not found: ID does not exist" containerID="f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.561714 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7"} err="failed to get container status \"f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7\": rpc error: code = NotFound desc = could not find container \"f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7\": container with ID starting with f0dbd3ed4049fff7ca3598f0d9087c35a3ea980191cfde25412535c84023c8a7 not found: ID does not exist" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.561726 4843 scope.go:117] "RemoveContainer" containerID="0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa" Mar 14 09:50:47 crc kubenswrapper[4843]: E0314 09:50:47.562174 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa\": container with ID starting with 0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa not found: ID does not exist" containerID="0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.562235 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa"} err="failed to get container status \"0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa\": rpc error: code = NotFound desc = could not find container \"0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa\": container with ID starting with 0291fd182a54223ea56d92426f920c16c76c0999149bf36c91e81eedb4cca8fa not found: ID does not exist" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671040 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pj4w\" (UniqueName: \"kubernetes.io/projected/332d9367-2426-4579-93cd-c96252af6090-kube-api-access-5pj4w\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671097 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-config-data\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671123 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-log-httpd\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671147 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671171 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-scripts\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671191 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671251 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-run-httpd\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.671317 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.773732 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.774227 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pj4w\" (UniqueName: \"kubernetes.io/projected/332d9367-2426-4579-93cd-c96252af6090-kube-api-access-5pj4w\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.774508 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-config-data\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.774709 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-log-httpd\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.774885 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.775077 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-scripts\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.775303 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.775614 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-run-httpd\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.776663 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-run-httpd\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.778733 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-log-httpd\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.783872 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.784665 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-config-data\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.785585 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.786169 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.801429 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-scripts\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.814118 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pj4w\" (UniqueName: \"kubernetes.io/projected/332d9367-2426-4579-93cd-c96252af6090-kube-api-access-5pj4w\") pod \"ceilometer-0\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:47 crc kubenswrapper[4843]: I0314 09:50:47.836710 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:48 crc kubenswrapper[4843]: I0314 09:50:48.231699 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:50:48 crc kubenswrapper[4843]: I0314 09:50:48.243167 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:50:48 crc kubenswrapper[4843]: I0314 09:50:48.459249 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerStarted","Data":"c71a5f939223b3e0ab68f94add5c5e4e5c7d1cb39d3a0de607ae12112f34dbaf"} Mar 14 09:50:49 crc kubenswrapper[4843]: I0314 09:50:49.350160 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28ca6ab9-a606-4631-8714-dbca98f6cfcd" path="/var/lib/kubelet/pods/28ca6ab9-a606-4631-8714-dbca98f6cfcd/volumes" Mar 14 09:50:49 crc kubenswrapper[4843]: I0314 09:50:49.469320 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerStarted","Data":"afdb4c8617528309845b42d2dbccbb9275fed12ae0586ad2c08b3ec68f467a31"} Mar 14 09:50:50 crc kubenswrapper[4843]: I0314 09:50:50.317211 4843 scope.go:117] "RemoveContainer" containerID="29cef80e012d63a5c0a221b4bf1dc5afb0078ca7589dbc0102184f7aaa699255" Mar 14 09:50:50 crc kubenswrapper[4843]: I0314 09:50:50.479318 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerStarted","Data":"0630c13543895bc2d1f4e511d3e745e8bbb62de1b1cee339f0e19496de0d8aee"} Mar 14 09:50:50 crc kubenswrapper[4843]: I0314 09:50:50.479361 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerStarted","Data":"1c6382b86c92b9d70e09cd01caa74c0fb3fad47108f970198fbe644e2922aae9"} Mar 14 09:50:52 crc kubenswrapper[4843]: I0314 09:50:52.507042 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerStarted","Data":"d9f8c36fbcef7e13f7b5aea054b0505e7e72074e8413cfd08b8d8d392a084f49"} Mar 14 09:50:52 crc kubenswrapper[4843]: I0314 09:50:52.508966 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:50:52 crc kubenswrapper[4843]: I0314 09:50:52.531287 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.054540148 podStartE2EDuration="5.531252075s" podCreationTimestamp="2026-03-14 09:50:47 +0000 UTC" firstStartedPulling="2026-03-14 09:50:48.242954094 +0000 UTC m=+2415.555565222" lastFinishedPulling="2026-03-14 09:50:51.719666011 +0000 UTC m=+2419.032277149" observedRunningTime="2026-03-14 09:50:52.526686903 +0000 UTC m=+2419.839298031" watchObservedRunningTime="2026-03-14 09:50:52.531252075 +0000 UTC m=+2419.843863203" Mar 14 09:50:56 crc kubenswrapper[4843]: I0314 09:50:56.340134 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:50:56 crc kubenswrapper[4843]: E0314 09:50:56.340951 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:51:01 crc kubenswrapper[4843]: E0314 09:51:01.363028 4843 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.162:38832->38.102.83.162:34757: read tcp 38.102.83.162:38832->38.102.83.162:34757: read: connection reset by peer Mar 14 09:51:10 crc kubenswrapper[4843]: I0314 09:51:10.338596 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:51:10 crc kubenswrapper[4843]: E0314 09:51:10.339220 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:51:17 crc kubenswrapper[4843]: I0314 09:51:17.845479 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.333008 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-x29w7"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.342922 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-x29w7"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.373146 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.373373 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="f31484f2-c78d-4896-ab8d-c6aa331c5255" containerName="watcher-decision-engine" containerID="cri-o://5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36" gracePeriod=30 Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.400457 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcherc17a-account-delete-7qsfb"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.401455 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.436954 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherc17a-account-delete-7qsfb"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.478396 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.478673 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-kuttl-api-log" containerID="cri-o://e7c063ed07886f5e11ea86ca7dcb680b4f6c864b7fd9965fec71ba1e890a85e4" gracePeriod=30 Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.479335 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-api" containerID="cri-o://0a82902b88a501520e4b9cadbb9703a08ea48cf796e2d422fbb8ea815958b146" gracePeriod=30 Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.507707 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.507927 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="a31aec38-dc3f-4143-8b4d-bc7b61f169ea" containerName="watcher-applier" containerID="cri-o://5626a70f5fca2153346a27da538352451986038107b6c35b5f5007104f430a8a" gracePeriod=30 Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.581311 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-operator-scripts\") pod \"watcherc17a-account-delete-7qsfb\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.581407 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdfxd\" (UniqueName: \"kubernetes.io/projected/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-kube-api-access-qdfxd\") pod \"watcherc17a-account-delete-7qsfb\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.682636 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-operator-scripts\") pod \"watcherc17a-account-delete-7qsfb\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.682733 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdfxd\" (UniqueName: \"kubernetes.io/projected/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-kube-api-access-qdfxd\") pod \"watcherc17a-account-delete-7qsfb\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.683542 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-operator-scripts\") pod \"watcherc17a-account-delete-7qsfb\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.714303 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdfxd\" (UniqueName: \"kubernetes.io/projected/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-kube-api-access-qdfxd\") pod \"watcherc17a-account-delete-7qsfb\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.737738 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.748159 4843 generic.go:334] "Generic (PLEG): container finished" podID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerID="e7c063ed07886f5e11ea86ca7dcb680b4f6c864b7fd9965fec71ba1e890a85e4" exitCode=143 Mar 14 09:51:22 crc kubenswrapper[4843]: I0314 09:51:22.748208 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7","Type":"ContainerDied","Data":"e7c063ed07886f5e11ea86ca7dcb680b4f6c864b7fd9965fec71ba1e890a85e4"} Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.253431 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherc17a-account-delete-7qsfb"] Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.344467 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:51:23 crc kubenswrapper[4843]: E0314 09:51:23.344951 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.347602 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="448d3cda-6240-4734-b55a-65a8d33a69f6" path="/var/lib/kubelet/pods/448d3cda-6240-4734-b55a-65a8d33a69f6/volumes" Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.530420 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.204:9322/\": dial tcp 10.217.0.204:9322: connect: connection refused" Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.530745 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.204:9322/\": dial tcp 10.217.0.204:9322: connect: connection refused" Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.773888 4843 generic.go:334] "Generic (PLEG): container finished" podID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerID="0a82902b88a501520e4b9cadbb9703a08ea48cf796e2d422fbb8ea815958b146" exitCode=0 Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.773942 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7","Type":"ContainerDied","Data":"0a82902b88a501520e4b9cadbb9703a08ea48cf796e2d422fbb8ea815958b146"} Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.777228 4843 generic.go:334] "Generic (PLEG): container finished" podID="a31aec38-dc3f-4143-8b4d-bc7b61f169ea" containerID="5626a70f5fca2153346a27da538352451986038107b6c35b5f5007104f430a8a" exitCode=0 Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.777310 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"a31aec38-dc3f-4143-8b4d-bc7b61f169ea","Type":"ContainerDied","Data":"5626a70f5fca2153346a27da538352451986038107b6c35b5f5007104f430a8a"} Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.779348 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" event={"ID":"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465","Type":"ContainerStarted","Data":"bc0ebf7333899aae0cbf31cfc8a0453b4462315bb665ab7383b5c06f1b08be90"} Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.779383 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" event={"ID":"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465","Type":"ContainerStarted","Data":"edd7f3c36a64d2b0eb9bdabb6eefb3943c0a5cb1af27a7a6b6130b7facb606c7"} Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.800509 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" podStartSLOduration=1.800488914 podStartE2EDuration="1.800488914s" podCreationTimestamp="2026-03-14 09:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:51:23.792888338 +0000 UTC m=+2451.105499466" watchObservedRunningTime="2026-03-14 09:51:23.800488914 +0000 UTC m=+2451.113100042" Mar 14 09:51:23 crc kubenswrapper[4843]: I0314 09:51:23.975598 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.050204 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-cert-memcached-mtls\") pod \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.050259 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-combined-ca-bundle\") pod \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.050360 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swkf5\" (UniqueName: \"kubernetes.io/projected/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-kube-api-access-swkf5\") pod \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.050413 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-config-data\") pod \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.050514 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-logs\") pod \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\" (UID: \"a31aec38-dc3f-4143-8b4d-bc7b61f169ea\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.051103 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-logs" (OuterVolumeSpecName: "logs") pod "a31aec38-dc3f-4143-8b4d-bc7b61f169ea" (UID: "a31aec38-dc3f-4143-8b4d-bc7b61f169ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.053854 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.058882 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-kube-api-access-swkf5" (OuterVolumeSpecName: "kube-api-access-swkf5") pod "a31aec38-dc3f-4143-8b4d-bc7b61f169ea" (UID: "a31aec38-dc3f-4143-8b4d-bc7b61f169ea"). InnerVolumeSpecName "kube-api-access-swkf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.085850 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a31aec38-dc3f-4143-8b4d-bc7b61f169ea" (UID: "a31aec38-dc3f-4143-8b4d-bc7b61f169ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.116412 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-config-data" (OuterVolumeSpecName: "config-data") pod "a31aec38-dc3f-4143-8b4d-bc7b61f169ea" (UID: "a31aec38-dc3f-4143-8b4d-bc7b61f169ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.143672 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "a31aec38-dc3f-4143-8b4d-bc7b61f169ea" (UID: "a31aec38-dc3f-4143-8b4d-bc7b61f169ea"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.151479 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-custom-prometheus-ca\") pod \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.151543 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7lvm\" (UniqueName: \"kubernetes.io/projected/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-kube-api-access-t7lvm\") pod \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.151701 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-logs\") pod \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.151732 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-config-data\") pod \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.151881 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-cert-memcached-mtls\") pod \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.151949 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-combined-ca-bundle\") pod \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\" (UID: \"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7\") " Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152151 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-logs" (OuterVolumeSpecName: "logs") pod "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" (UID: "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152535 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152578 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152592 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152603 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152613 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swkf5\" (UniqueName: \"kubernetes.io/projected/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-kube-api-access-swkf5\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.152644 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31aec38-dc3f-4143-8b4d-bc7b61f169ea-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.154570 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-kube-api-access-t7lvm" (OuterVolumeSpecName: "kube-api-access-t7lvm") pod "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" (UID: "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7"). InnerVolumeSpecName "kube-api-access-t7lvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.175522 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" (UID: "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.178525 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" (UID: "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.198534 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-config-data" (OuterVolumeSpecName: "config-data") pod "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" (UID: "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.233630 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" (UID: "58112b9d-0e99-4e9b-9b6a-655c1ba03bf7"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.255132 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.255179 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.255205 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.255219 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.255235 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7lvm\" (UniqueName: \"kubernetes.io/projected/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7-kube-api-access-t7lvm\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.787871 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"a31aec38-dc3f-4143-8b4d-bc7b61f169ea","Type":"ContainerDied","Data":"dfa87d670635984792f5471ad13f3d93cc15e7cf3e91f6b25332510e57315dd9"} Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.788203 4843 scope.go:117] "RemoveContainer" containerID="5626a70f5fca2153346a27da538352451986038107b6c35b5f5007104f430a8a" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.787880 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.790829 4843 generic.go:334] "Generic (PLEG): container finished" podID="3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" containerID="bc0ebf7333899aae0cbf31cfc8a0453b4462315bb665ab7383b5c06f1b08be90" exitCode=0 Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.790984 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" event={"ID":"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465","Type":"ContainerDied","Data":"bc0ebf7333899aae0cbf31cfc8a0453b4462315bb665ab7383b5c06f1b08be90"} Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.794258 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"58112b9d-0e99-4e9b-9b6a-655c1ba03bf7","Type":"ContainerDied","Data":"0128996aa9a8f9be820eb9a3d2e8cbc5ad05be152f0eba804ae52c8b79b24c36"} Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.794543 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.808890 4843 scope.go:117] "RemoveContainer" containerID="0a82902b88a501520e4b9cadbb9703a08ea48cf796e2d422fbb8ea815958b146" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.835573 4843 scope.go:117] "RemoveContainer" containerID="e7c063ed07886f5e11ea86ca7dcb680b4f6c864b7fd9965fec71ba1e890a85e4" Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.844579 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.853457 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.860463 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.867049 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.971811 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.972066 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-central-agent" containerID="cri-o://afdb4c8617528309845b42d2dbccbb9275fed12ae0586ad2c08b3ec68f467a31" gracePeriod=30 Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.972158 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-notification-agent" containerID="cri-o://1c6382b86c92b9d70e09cd01caa74c0fb3fad47108f970198fbe644e2922aae9" gracePeriod=30 Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.972190 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="proxy-httpd" containerID="cri-o://d9f8c36fbcef7e13f7b5aea054b0505e7e72074e8413cfd08b8d8d392a084f49" gracePeriod=30 Mar 14 09:51:24 crc kubenswrapper[4843]: I0314 09:51:24.972552 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="sg-core" containerID="cri-o://0630c13543895bc2d1f4e511d3e745e8bbb62de1b1cee339f0e19496de0d8aee" gracePeriod=30 Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.346998 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" path="/var/lib/kubelet/pods/58112b9d-0e99-4e9b-9b6a-655c1ba03bf7/volumes" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.347625 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31aec38-dc3f-4143-8b4d-bc7b61f169ea" path="/var/lib/kubelet/pods/a31aec38-dc3f-4143-8b4d-bc7b61f169ea/volumes" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.808790 4843 generic.go:334] "Generic (PLEG): container finished" podID="332d9367-2426-4579-93cd-c96252af6090" containerID="d9f8c36fbcef7e13f7b5aea054b0505e7e72074e8413cfd08b8d8d392a084f49" exitCode=0 Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809099 4843 generic.go:334] "Generic (PLEG): container finished" podID="332d9367-2426-4579-93cd-c96252af6090" containerID="0630c13543895bc2d1f4e511d3e745e8bbb62de1b1cee339f0e19496de0d8aee" exitCode=2 Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.808852 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerDied","Data":"d9f8c36fbcef7e13f7b5aea054b0505e7e72074e8413cfd08b8d8d392a084f49"} Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809141 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerDied","Data":"0630c13543895bc2d1f4e511d3e745e8bbb62de1b1cee339f0e19496de0d8aee"} Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809157 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerDied","Data":"1c6382b86c92b9d70e09cd01caa74c0fb3fad47108f970198fbe644e2922aae9"} Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809112 4843 generic.go:334] "Generic (PLEG): container finished" podID="332d9367-2426-4579-93cd-c96252af6090" containerID="1c6382b86c92b9d70e09cd01caa74c0fb3fad47108f970198fbe644e2922aae9" exitCode=0 Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809177 4843 generic.go:334] "Generic (PLEG): container finished" podID="332d9367-2426-4579-93cd-c96252af6090" containerID="afdb4c8617528309845b42d2dbccbb9275fed12ae0586ad2c08b3ec68f467a31" exitCode=0 Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809234 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerDied","Data":"afdb4c8617528309845b42d2dbccbb9275fed12ae0586ad2c08b3ec68f467a31"} Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809306 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"332d9367-2426-4579-93cd-c96252af6090","Type":"ContainerDied","Data":"c71a5f939223b3e0ab68f94add5c5e4e5c7d1cb39d3a0de607ae12112f34dbaf"} Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.809317 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c71a5f939223b3e0ab68f94add5c5e4e5c7d1cb39d3a0de607ae12112f34dbaf" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.853134 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.906658 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-ceilometer-tls-certs\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.906719 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-scripts\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.906789 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pj4w\" (UniqueName: \"kubernetes.io/projected/332d9367-2426-4579-93cd-c96252af6090-kube-api-access-5pj4w\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.906829 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-sg-core-conf-yaml\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.906861 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-log-httpd\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.906999 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-run-httpd\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.907025 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-combined-ca-bundle\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.907101 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-config-data\") pod \"332d9367-2426-4579-93cd-c96252af6090\" (UID: \"332d9367-2426-4579-93cd-c96252af6090\") " Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.909083 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.909704 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.917111 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-scripts" (OuterVolumeSpecName: "scripts") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.918454 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332d9367-2426-4579-93cd-c96252af6090-kube-api-access-5pj4w" (OuterVolumeSpecName: "kube-api-access-5pj4w") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "kube-api-access-5pj4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.945407 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:25 crc kubenswrapper[4843]: I0314 09:51:25.980066 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.011237 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.011272 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.011295 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pj4w\" (UniqueName: \"kubernetes.io/projected/332d9367-2426-4579-93cd-c96252af6090-kube-api-access-5pj4w\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.011305 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.011313 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.011320 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/332d9367-2426-4579-93cd-c96252af6090-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.046745 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-config-data" (OuterVolumeSpecName: "config-data") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.049932 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "332d9367-2426-4579-93cd-c96252af6090" (UID: "332d9367-2426-4579-93cd-c96252af6090"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.113072 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.113111 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d9367-2426-4579-93cd-c96252af6090-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.225117 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.316790 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-operator-scripts\") pod \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.317023 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdfxd\" (UniqueName: \"kubernetes.io/projected/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-kube-api-access-qdfxd\") pod \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\" (UID: \"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.318732 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" (UID: "3a5b47ed-0f7a-4d4f-8c9d-c30d63803465"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.320474 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-kube-api-access-qdfxd" (OuterVolumeSpecName: "kube-api-access-qdfxd") pod "3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" (UID: "3a5b47ed-0f7a-4d4f-8c9d-c30d63803465"). InnerVolumeSpecName "kube-api-access-qdfxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.364754 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.426820 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8brn\" (UniqueName: \"kubernetes.io/projected/f31484f2-c78d-4896-ab8d-c6aa331c5255-kube-api-access-t8brn\") pod \"f31484f2-c78d-4896-ab8d-c6aa331c5255\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.426882 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-custom-prometheus-ca\") pod \"f31484f2-c78d-4896-ab8d-c6aa331c5255\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.426979 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-cert-memcached-mtls\") pod \"f31484f2-c78d-4896-ab8d-c6aa331c5255\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.427010 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-combined-ca-bundle\") pod \"f31484f2-c78d-4896-ab8d-c6aa331c5255\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.427046 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f31484f2-c78d-4896-ab8d-c6aa331c5255-logs\") pod \"f31484f2-c78d-4896-ab8d-c6aa331c5255\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.427113 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-config-data\") pod \"f31484f2-c78d-4896-ab8d-c6aa331c5255\" (UID: \"f31484f2-c78d-4896-ab8d-c6aa331c5255\") " Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.427542 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdfxd\" (UniqueName: \"kubernetes.io/projected/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-kube-api-access-qdfxd\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.427567 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.428585 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f31484f2-c78d-4896-ab8d-c6aa331c5255-logs" (OuterVolumeSpecName: "logs") pod "f31484f2-c78d-4896-ab8d-c6aa331c5255" (UID: "f31484f2-c78d-4896-ab8d-c6aa331c5255"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.437446 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f31484f2-c78d-4896-ab8d-c6aa331c5255-kube-api-access-t8brn" (OuterVolumeSpecName: "kube-api-access-t8brn") pod "f31484f2-c78d-4896-ab8d-c6aa331c5255" (UID: "f31484f2-c78d-4896-ab8d-c6aa331c5255"). InnerVolumeSpecName "kube-api-access-t8brn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.461999 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f31484f2-c78d-4896-ab8d-c6aa331c5255" (UID: "f31484f2-c78d-4896-ab8d-c6aa331c5255"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.465091 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "f31484f2-c78d-4896-ab8d-c6aa331c5255" (UID: "f31484f2-c78d-4896-ab8d-c6aa331c5255"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.495208 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-config-data" (OuterVolumeSpecName: "config-data") pod "f31484f2-c78d-4896-ab8d-c6aa331c5255" (UID: "f31484f2-c78d-4896-ab8d-c6aa331c5255"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.505738 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "f31484f2-c78d-4896-ab8d-c6aa331c5255" (UID: "f31484f2-c78d-4896-ab8d-c6aa331c5255"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.528792 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.528826 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.528835 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f31484f2-c78d-4896-ab8d-c6aa331c5255-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.528845 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.528855 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8brn\" (UniqueName: \"kubernetes.io/projected/f31484f2-c78d-4896-ab8d-c6aa331c5255-kube-api-access-t8brn\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.528866 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f31484f2-c78d-4896-ab8d-c6aa331c5255-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.819022 4843 generic.go:334] "Generic (PLEG): container finished" podID="f31484f2-c78d-4896-ab8d-c6aa331c5255" containerID="5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36" exitCode=0 Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.819094 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"f31484f2-c78d-4896-ab8d-c6aa331c5255","Type":"ContainerDied","Data":"5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36"} Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.819107 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.819136 4843 scope.go:117] "RemoveContainer" containerID="5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.819124 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"f31484f2-c78d-4896-ab8d-c6aa331c5255","Type":"ContainerDied","Data":"cb843b526d8c52e8e60f6f72c0219ae7f394e00e2706aa4ed78adf2daca0f095"} Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.823092 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" event={"ID":"3a5b47ed-0f7a-4d4f-8c9d-c30d63803465","Type":"ContainerDied","Data":"edd7f3c36a64d2b0eb9bdabb6eefb3943c0a5cb1af27a7a6b6130b7facb606c7"} Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.823136 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edd7f3c36a64d2b0eb9bdabb6eefb3943c0a5cb1af27a7a6b6130b7facb606c7" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.823143 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.823204 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherc17a-account-delete-7qsfb" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.884903 4843 scope.go:117] "RemoveContainer" containerID="5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.885377 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36\": container with ID starting with 5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36 not found: ID does not exist" containerID="5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.885418 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36"} err="failed to get container status \"5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36\": rpc error: code = NotFound desc = could not find container \"5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36\": container with ID starting with 5b6f3c8d6bbd87e13753e7017f760d87a1f5c525c34dcb83c72438a190aaea36 not found: ID does not exist" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.896779 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.911998 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.921397 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.928041 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.947959 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948444 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-kuttl-api-log" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948468 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-kuttl-api-log" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948491 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a31aec38-dc3f-4143-8b4d-bc7b61f169ea" containerName="watcher-applier" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948500 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="a31aec38-dc3f-4143-8b4d-bc7b61f169ea" containerName="watcher-applier" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948518 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f31484f2-c78d-4896-ab8d-c6aa331c5255" containerName="watcher-decision-engine" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948525 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f31484f2-c78d-4896-ab8d-c6aa331c5255" containerName="watcher-decision-engine" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948535 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-api" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948542 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-api" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948553 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-central-agent" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948561 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-central-agent" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948569 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="proxy-httpd" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948576 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="proxy-httpd" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948591 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="sg-core" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948598 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="sg-core" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948606 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" containerName="mariadb-account-delete" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948613 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" containerName="mariadb-account-delete" Mar 14 09:51:26 crc kubenswrapper[4843]: E0314 09:51:26.948624 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-notification-agent" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948631 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-notification-agent" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948788 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-notification-agent" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948803 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" containerName="mariadb-account-delete" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948817 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="sg-core" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948829 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-api" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948838 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="ceilometer-central-agent" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948848 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="332d9367-2426-4579-93cd-c96252af6090" containerName="proxy-httpd" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948859 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="a31aec38-dc3f-4143-8b4d-bc7b61f169ea" containerName="watcher-applier" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948870 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f31484f2-c78d-4896-ab8d-c6aa331c5255" containerName="watcher-decision-engine" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.948881 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="58112b9d-0e99-4e9b-9b6a-655c1ba03bf7" containerName="watcher-kuttl-api-log" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.950405 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.952890 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.952998 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.953477 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:51:26 crc kubenswrapper[4843]: I0314 09:51:26.955535 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037174 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-config-data\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037497 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-run-httpd\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037525 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n67np\" (UniqueName: \"kubernetes.io/projected/441d7f37-0bad-4a93-94f6-71d380ca73b1-kube-api-access-n67np\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037575 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037600 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037634 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-log-httpd\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037662 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-scripts\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.037686 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138814 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-config-data\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138866 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-run-httpd\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138892 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n67np\" (UniqueName: \"kubernetes.io/projected/441d7f37-0bad-4a93-94f6-71d380ca73b1-kube-api-access-n67np\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138932 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138954 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138971 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-log-httpd\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.138995 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-scripts\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.139016 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.139652 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-log-httpd\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.140196 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-run-httpd\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.144304 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-scripts\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.145374 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-config-data\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.145980 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.148037 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.153123 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.184891 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n67np\" (UniqueName: \"kubernetes.io/projected/441d7f37-0bad-4a93-94f6-71d380ca73b1-kube-api-access-n67np\") pod \"ceilometer-0\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.266565 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.352768 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332d9367-2426-4579-93cd-c96252af6090" path="/var/lib/kubelet/pods/332d9367-2426-4579-93cd-c96252af6090/volumes" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.353927 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f31484f2-c78d-4896-ab8d-c6aa331c5255" path="/var/lib/kubelet/pods/f31484f2-c78d-4896-ab8d-c6aa331c5255/volumes" Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.454462 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5tng5"] Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.488586 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-5tng5"] Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.502521 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5"] Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.509109 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcherc17a-account-delete-7qsfb"] Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.515376 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-c17a-account-create-update-jtsj5"] Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.530822 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcherc17a-account-delete-7qsfb"] Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.797386 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:27 crc kubenswrapper[4843]: W0314 09:51:27.802822 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod441d7f37_0bad_4a93_94f6_71d380ca73b1.slice/crio-2360222e37d4cec9d7419bf07885505b14c52ed2e032352b086885b94d94deb4 WatchSource:0}: Error finding container 2360222e37d4cec9d7419bf07885505b14c52ed2e032352b086885b94d94deb4: Status 404 returned error can't find the container with id 2360222e37d4cec9d7419bf07885505b14c52ed2e032352b086885b94d94deb4 Mar 14 09:51:27 crc kubenswrapper[4843]: I0314 09:51:27.831493 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerStarted","Data":"2360222e37d4cec9d7419bf07885505b14c52ed2e032352b086885b94d94deb4"} Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.841238 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerStarted","Data":"cda2091bf4f3e6e0e5a74f0d2259fdde47b20124b53fd77d8fa80662968fddbc"} Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.853317 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-nx2zd"] Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.854659 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.860694 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp"] Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.861932 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.864480 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.870946 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-nx2zd"] Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.881957 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp"] Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.967638 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/959c17a9-c974-49e8-bf5a-60be6bf6a944-operator-scripts\") pod \"watcher-5b3a-account-create-update-rlchp\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.968041 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-operator-scripts\") pod \"watcher-db-create-nx2zd\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.968187 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl2kd\" (UniqueName: \"kubernetes.io/projected/959c17a9-c974-49e8-bf5a-60be6bf6a944-kube-api-access-fl2kd\") pod \"watcher-5b3a-account-create-update-rlchp\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:28 crc kubenswrapper[4843]: I0314 09:51:28.968321 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjb8r\" (UniqueName: \"kubernetes.io/projected/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-kube-api-access-kjb8r\") pod \"watcher-db-create-nx2zd\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.070548 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl2kd\" (UniqueName: \"kubernetes.io/projected/959c17a9-c974-49e8-bf5a-60be6bf6a944-kube-api-access-fl2kd\") pod \"watcher-5b3a-account-create-update-rlchp\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.070579 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjb8r\" (UniqueName: \"kubernetes.io/projected/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-kube-api-access-kjb8r\") pod \"watcher-db-create-nx2zd\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.070639 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/959c17a9-c974-49e8-bf5a-60be6bf6a944-operator-scripts\") pod \"watcher-5b3a-account-create-update-rlchp\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.070689 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-operator-scripts\") pod \"watcher-db-create-nx2zd\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.071804 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-operator-scripts\") pod \"watcher-db-create-nx2zd\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.071853 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/959c17a9-c974-49e8-bf5a-60be6bf6a944-operator-scripts\") pod \"watcher-5b3a-account-create-update-rlchp\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.091213 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjb8r\" (UniqueName: \"kubernetes.io/projected/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-kube-api-access-kjb8r\") pod \"watcher-db-create-nx2zd\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.093169 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl2kd\" (UniqueName: \"kubernetes.io/projected/959c17a9-c974-49e8-bf5a-60be6bf6a944-kube-api-access-fl2kd\") pod \"watcher-5b3a-account-create-update-rlchp\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.179838 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.333740 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.360017 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f55d604-7f52-420f-ac0c-7d3e68a2bffc" path="/var/lib/kubelet/pods/2f55d604-7f52-420f-ac0c-7d3e68a2bffc/volumes" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.361086 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a5b47ed-0f7a-4d4f-8c9d-c30d63803465" path="/var/lib/kubelet/pods/3a5b47ed-0f7a-4d4f-8c9d-c30d63803465/volumes" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.361978 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9" path="/var/lib/kubelet/pods/8c780c8c-e7e7-41b8-adc6-0c8726e1d7f9/volumes" Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.745337 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-nx2zd"] Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.852055 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerStarted","Data":"3cbd0922fb7c2a426e4a6f1b87cfca5ed741f1b809766b207df4950d48efdacf"} Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.852086 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerStarted","Data":"e478bfbbf9b07eb8b1c597b506f27d645773dd7382092926e5d126c7cd23b643"} Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.858015 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-nx2zd" event={"ID":"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f","Type":"ContainerStarted","Data":"59296e1d170a85e5e724fd7c1ac7b7817ac8aa2a75b260e5aeaaca2e3dba2aea"} Mar 14 09:51:29 crc kubenswrapper[4843]: I0314 09:51:29.899104 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp"] Mar 14 09:51:30 crc kubenswrapper[4843]: I0314 09:51:30.865756 4843 generic.go:334] "Generic (PLEG): container finished" podID="959c17a9-c974-49e8-bf5a-60be6bf6a944" containerID="50851d18785e26c731d169912ceb516dd101c013eebe32bd01d97b47ded00d0a" exitCode=0 Mar 14 09:51:30 crc kubenswrapper[4843]: I0314 09:51:30.865819 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" event={"ID":"959c17a9-c974-49e8-bf5a-60be6bf6a944","Type":"ContainerDied","Data":"50851d18785e26c731d169912ceb516dd101c013eebe32bd01d97b47ded00d0a"} Mar 14 09:51:30 crc kubenswrapper[4843]: I0314 09:51:30.866079 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" event={"ID":"959c17a9-c974-49e8-bf5a-60be6bf6a944","Type":"ContainerStarted","Data":"e7e32bafe1630eb997b26e4e84e388fa6437de01dbea24e59d55a47b35ec8967"} Mar 14 09:51:30 crc kubenswrapper[4843]: I0314 09:51:30.868745 4843 generic.go:334] "Generic (PLEG): container finished" podID="1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" containerID="3c6e0df8928364da3119289d1ba8b72295a281e8e99c8f0fefc0ea531db15869" exitCode=0 Mar 14 09:51:30 crc kubenswrapper[4843]: I0314 09:51:30.868799 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-nx2zd" event={"ID":"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f","Type":"ContainerDied","Data":"3c6e0df8928364da3119289d1ba8b72295a281e8e99c8f0fefc0ea531db15869"} Mar 14 09:51:31 crc kubenswrapper[4843]: I0314 09:51:31.880486 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerStarted","Data":"419165322e603cd53c2137dbb1d5630a01f35bf58e5bb16bb782c5da7958987e"} Mar 14 09:51:31 crc kubenswrapper[4843]: I0314 09:51:31.912567 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.5801536069999997 podStartE2EDuration="5.912525058s" podCreationTimestamp="2026-03-14 09:51:26 +0000 UTC" firstStartedPulling="2026-03-14 09:51:27.80600478 +0000 UTC m=+2455.118615918" lastFinishedPulling="2026-03-14 09:51:31.138376241 +0000 UTC m=+2458.450987369" observedRunningTime="2026-03-14 09:51:31.902614926 +0000 UTC m=+2459.215226054" watchObservedRunningTime="2026-03-14 09:51:31.912525058 +0000 UTC m=+2459.225136186" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.338646 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.426552 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-operator-scripts\") pod \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.426650 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjb8r\" (UniqueName: \"kubernetes.io/projected/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-kube-api-access-kjb8r\") pod \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\" (UID: \"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f\") " Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.427116 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" (UID: "1fd8d7c1-b35b-4982-9cf4-04ed8a29192f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.433071 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-kube-api-access-kjb8r" (OuterVolumeSpecName: "kube-api-access-kjb8r") pod "1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" (UID: "1fd8d7c1-b35b-4982-9cf4-04ed8a29192f"). InnerVolumeSpecName "kube-api-access-kjb8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.466627 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.527999 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/959c17a9-c974-49e8-bf5a-60be6bf6a944-operator-scripts\") pod \"959c17a9-c974-49e8-bf5a-60be6bf6a944\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.528041 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl2kd\" (UniqueName: \"kubernetes.io/projected/959c17a9-c974-49e8-bf5a-60be6bf6a944-kube-api-access-fl2kd\") pod \"959c17a9-c974-49e8-bf5a-60be6bf6a944\" (UID: \"959c17a9-c974-49e8-bf5a-60be6bf6a944\") " Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.528515 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.528534 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjb8r\" (UniqueName: \"kubernetes.io/projected/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f-kube-api-access-kjb8r\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.528543 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/959c17a9-c974-49e8-bf5a-60be6bf6a944-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "959c17a9-c974-49e8-bf5a-60be6bf6a944" (UID: "959c17a9-c974-49e8-bf5a-60be6bf6a944"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.531016 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/959c17a9-c974-49e8-bf5a-60be6bf6a944-kube-api-access-fl2kd" (OuterVolumeSpecName: "kube-api-access-fl2kd") pod "959c17a9-c974-49e8-bf5a-60be6bf6a944" (UID: "959c17a9-c974-49e8-bf5a-60be6bf6a944"). InnerVolumeSpecName "kube-api-access-fl2kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.630208 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/959c17a9-c974-49e8-bf5a-60be6bf6a944-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.630235 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl2kd\" (UniqueName: \"kubernetes.io/projected/959c17a9-c974-49e8-bf5a-60be6bf6a944-kube-api-access-fl2kd\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.904618 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" event={"ID":"959c17a9-c974-49e8-bf5a-60be6bf6a944","Type":"ContainerDied","Data":"e7e32bafe1630eb997b26e4e84e388fa6437de01dbea24e59d55a47b35ec8967"} Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.904689 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7e32bafe1630eb997b26e4e84e388fa6437de01dbea24e59d55a47b35ec8967" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.904779 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.908237 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-nx2zd" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.908291 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-nx2zd" event={"ID":"1fd8d7c1-b35b-4982-9cf4-04ed8a29192f","Type":"ContainerDied","Data":"59296e1d170a85e5e724fd7c1ac7b7817ac8aa2a75b260e5aeaaca2e3dba2aea"} Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.908321 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59296e1d170a85e5e724fd7c1ac7b7817ac8aa2a75b260e5aeaaca2e3dba2aea" Mar 14 09:51:32 crc kubenswrapper[4843]: I0314 09:51:32.908422 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.564857 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8"] Mar 14 09:51:34 crc kubenswrapper[4843]: E0314 09:51:34.565463 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" containerName="mariadb-database-create" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.565477 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" containerName="mariadb-database-create" Mar 14 09:51:34 crc kubenswrapper[4843]: E0314 09:51:34.565507 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959c17a9-c974-49e8-bf5a-60be6bf6a944" containerName="mariadb-account-create-update" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.565513 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="959c17a9-c974-49e8-bf5a-60be6bf6a944" containerName="mariadb-account-create-update" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.565654 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="959c17a9-c974-49e8-bf5a-60be6bf6a944" containerName="mariadb-account-create-update" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.565673 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" containerName="mariadb-database-create" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.566229 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.569309 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8"] Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.598069 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.599060 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-48jmf" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.659646 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm6x8\" (UniqueName: \"kubernetes.io/projected/f097b765-f4be-4e04-a4c9-8391e58aaf20-kube-api-access-wm6x8\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.659750 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-config-data\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.659779 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.659802 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-db-sync-config-data\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.761680 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm6x8\" (UniqueName: \"kubernetes.io/projected/f097b765-f4be-4e04-a4c9-8391e58aaf20-kube-api-access-wm6x8\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.761822 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-config-data\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.761857 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.761889 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-db-sync-config-data\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.774659 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-db-sync-config-data\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.774844 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.774926 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-config-data\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.779451 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm6x8\" (UniqueName: \"kubernetes.io/projected/f097b765-f4be-4e04-a4c9-8391e58aaf20-kube-api-access-wm6x8\") pod \"watcher-kuttl-db-sync-sk4w8\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:34 crc kubenswrapper[4843]: I0314 09:51:34.927070 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:35 crc kubenswrapper[4843]: I0314 09:51:35.339531 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:51:35 crc kubenswrapper[4843]: E0314 09:51:35.339978 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:51:35 crc kubenswrapper[4843]: I0314 09:51:35.639204 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8"] Mar 14 09:51:35 crc kubenswrapper[4843]: I0314 09:51:35.936734 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" event={"ID":"f097b765-f4be-4e04-a4c9-8391e58aaf20","Type":"ContainerStarted","Data":"4a2a26597a24034cd40558c7ee55ba24a97b59c22547df9bdf88068307b14065"} Mar 14 09:51:35 crc kubenswrapper[4843]: I0314 09:51:35.937002 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" event={"ID":"f097b765-f4be-4e04-a4c9-8391e58aaf20","Type":"ContainerStarted","Data":"6fbaa47eda1004f6588a5d64e03b68590f9d97c7457061694cee5a810c39e196"} Mar 14 09:51:38 crc kubenswrapper[4843]: I0314 09:51:38.972637 4843 generic.go:334] "Generic (PLEG): container finished" podID="f097b765-f4be-4e04-a4c9-8391e58aaf20" containerID="4a2a26597a24034cd40558c7ee55ba24a97b59c22547df9bdf88068307b14065" exitCode=0 Mar 14 09:51:38 crc kubenswrapper[4843]: I0314 09:51:38.972798 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" event={"ID":"f097b765-f4be-4e04-a4c9-8391e58aaf20","Type":"ContainerDied","Data":"4a2a26597a24034cd40558c7ee55ba24a97b59c22547df9bdf88068307b14065"} Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.343207 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.459232 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-db-sync-config-data\") pod \"f097b765-f4be-4e04-a4c9-8391e58aaf20\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.461068 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-config-data\") pod \"f097b765-f4be-4e04-a4c9-8391e58aaf20\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.461187 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-combined-ca-bundle\") pod \"f097b765-f4be-4e04-a4c9-8391e58aaf20\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.461263 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm6x8\" (UniqueName: \"kubernetes.io/projected/f097b765-f4be-4e04-a4c9-8391e58aaf20-kube-api-access-wm6x8\") pod \"f097b765-f4be-4e04-a4c9-8391e58aaf20\" (UID: \"f097b765-f4be-4e04-a4c9-8391e58aaf20\") " Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.464659 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f097b765-f4be-4e04-a4c9-8391e58aaf20" (UID: "f097b765-f4be-4e04-a4c9-8391e58aaf20"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.465014 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f097b765-f4be-4e04-a4c9-8391e58aaf20-kube-api-access-wm6x8" (OuterVolumeSpecName: "kube-api-access-wm6x8") pod "f097b765-f4be-4e04-a4c9-8391e58aaf20" (UID: "f097b765-f4be-4e04-a4c9-8391e58aaf20"). InnerVolumeSpecName "kube-api-access-wm6x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.490802 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f097b765-f4be-4e04-a4c9-8391e58aaf20" (UID: "f097b765-f4be-4e04-a4c9-8391e58aaf20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.501162 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-config-data" (OuterVolumeSpecName: "config-data") pod "f097b765-f4be-4e04-a4c9-8391e58aaf20" (UID: "f097b765-f4be-4e04-a4c9-8391e58aaf20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.563982 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.564019 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.564031 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm6x8\" (UniqueName: \"kubernetes.io/projected/f097b765-f4be-4e04-a4c9-8391e58aaf20-kube-api-access-wm6x8\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.564041 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f097b765-f4be-4e04-a4c9-8391e58aaf20-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.991574 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" event={"ID":"f097b765-f4be-4e04-a4c9-8391e58aaf20","Type":"ContainerDied","Data":"6fbaa47eda1004f6588a5d64e03b68590f9d97c7457061694cee5a810c39e196"} Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.991616 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fbaa47eda1004f6588a5d64e03b68590f9d97c7457061694cee5a810c39e196" Mar 14 09:51:40 crc kubenswrapper[4843]: I0314 09:51:40.991646 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.249094 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:41 crc kubenswrapper[4843]: E0314 09:51:41.249521 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f097b765-f4be-4e04-a4c9-8391e58aaf20" containerName="watcher-kuttl-db-sync" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.249544 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f097b765-f4be-4e04-a4c9-8391e58aaf20" containerName="watcher-kuttl-db-sync" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.249745 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f097b765-f4be-4e04-a4c9-8391e58aaf20" containerName="watcher-kuttl-db-sync" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.250842 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.252903 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.255388 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-48jmf" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.263559 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.273553 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274183 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274236 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274469 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274502 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274581 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbftp\" (UniqueName: \"kubernetes.io/projected/f73f2e72-f036-411d-9e88-c52f5795ec97-kube-api-access-zbftp\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274615 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.274656 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f73f2e72-f036-411d-9e88-c52f5795ec97-logs\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.277010 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.291165 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.367837 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.369168 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.372524 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.374895 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376143 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376210 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376252 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376311 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bba9218c-c199-4a70-bd95-f1c1de972639-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376333 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376356 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkjnf\" (UniqueName: \"kubernetes.io/projected/bba9218c-c199-4a70-bd95-f1c1de972639-kube-api-access-rkjnf\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376394 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376421 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376449 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b226653-0262-4cdb-9e8a-d3681331d87c-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376479 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbftp\" (UniqueName: \"kubernetes.io/projected/f73f2e72-f036-411d-9e88-c52f5795ec97-kube-api-access-zbftp\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376501 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwq28\" (UniqueName: \"kubernetes.io/projected/9b226653-0262-4cdb-9e8a-d3681331d87c-kube-api-access-gwq28\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376524 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376541 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f73f2e72-f036-411d-9e88-c52f5795ec97-logs\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376583 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376604 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376625 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.376646 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.380343 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.381377 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.382583 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f73f2e72-f036-411d-9e88-c52f5795ec97-logs\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.382643 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.388594 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.401316 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbftp\" (UniqueName: \"kubernetes.io/projected/f73f2e72-f036-411d-9e88-c52f5795ec97-kube-api-access-zbftp\") pod \"watcher-kuttl-api-0\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478474 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwq28\" (UniqueName: \"kubernetes.io/projected/9b226653-0262-4cdb-9e8a-d3681331d87c-kube-api-access-gwq28\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478558 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478583 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478606 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478636 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478684 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478717 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bba9218c-c199-4a70-bd95-f1c1de972639-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478734 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478749 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkjnf\" (UniqueName: \"kubernetes.io/projected/bba9218c-c199-4a70-bd95-f1c1de972639-kube-api-access-rkjnf\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478782 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.478799 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b226653-0262-4cdb-9e8a-d3681331d87c-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.479220 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b226653-0262-4cdb-9e8a-d3681331d87c-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.479241 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bba9218c-c199-4a70-bd95-f1c1de972639-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.482507 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.482618 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.482665 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.482696 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.483299 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.483807 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.486110 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.494299 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwq28\" (UniqueName: \"kubernetes.io/projected/9b226653-0262-4cdb-9e8a-d3681331d87c-kube-api-access-gwq28\") pod \"watcher-kuttl-applier-0\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.496644 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkjnf\" (UniqueName: \"kubernetes.io/projected/bba9218c-c199-4a70-bd95-f1c1de972639-kube-api-access-rkjnf\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.573715 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.590046 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:41 crc kubenswrapper[4843]: I0314 09:51:41.764045 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:42 crc kubenswrapper[4843]: I0314 09:51:42.219163 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:42 crc kubenswrapper[4843]: W0314 09:51:42.222476 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf73f2e72_f036_411d_9e88_c52f5795ec97.slice/crio-6371eeeade059315881a6873ef08f350a585471674b22de7d7ab289ccbc2e7f0 WatchSource:0}: Error finding container 6371eeeade059315881a6873ef08f350a585471674b22de7d7ab289ccbc2e7f0: Status 404 returned error can't find the container with id 6371eeeade059315881a6873ef08f350a585471674b22de7d7ab289ccbc2e7f0 Mar 14 09:51:42 crc kubenswrapper[4843]: I0314 09:51:42.301295 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:42 crc kubenswrapper[4843]: W0314 09:51:42.303922 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b226653_0262_4cdb_9e8a_d3681331d87c.slice/crio-cc9244c998471e557a992640588c115509e31a3effeb76c0f25ee6aa4c710960 WatchSource:0}: Error finding container cc9244c998471e557a992640588c115509e31a3effeb76c0f25ee6aa4c710960: Status 404 returned error can't find the container with id cc9244c998471e557a992640588c115509e31a3effeb76c0f25ee6aa4c710960 Mar 14 09:51:42 crc kubenswrapper[4843]: I0314 09:51:42.394548 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.010106 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"f73f2e72-f036-411d-9e88-c52f5795ec97","Type":"ContainerStarted","Data":"5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.010157 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"f73f2e72-f036-411d-9e88-c52f5795ec97","Type":"ContainerStarted","Data":"ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.010171 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"f73f2e72-f036-411d-9e88-c52f5795ec97","Type":"ContainerStarted","Data":"6371eeeade059315881a6873ef08f350a585471674b22de7d7ab289ccbc2e7f0"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.013294 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.015508 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"9b226653-0262-4cdb-9e8a-d3681331d87c","Type":"ContainerStarted","Data":"2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.015549 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"9b226653-0262-4cdb-9e8a-d3681331d87c","Type":"ContainerStarted","Data":"cc9244c998471e557a992640588c115509e31a3effeb76c0f25ee6aa4c710960"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.017910 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"bba9218c-c199-4a70-bd95-f1c1de972639","Type":"ContainerStarted","Data":"1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.017940 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"bba9218c-c199-4a70-bd95-f1c1de972639","Type":"ContainerStarted","Data":"8654108872b95a523010e371ccc79e2717f140ef7d8f597ee8d706afcaf019bc"} Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.039002 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.038980737 podStartE2EDuration="2.038980737s" podCreationTimestamp="2026-03-14 09:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:51:43.033552285 +0000 UTC m=+2470.346163413" watchObservedRunningTime="2026-03-14 09:51:43.038980737 +0000 UTC m=+2470.351591865" Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.056553 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.056533628 podStartE2EDuration="2.056533628s" podCreationTimestamp="2026-03-14 09:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:51:43.049719281 +0000 UTC m=+2470.362330429" watchObservedRunningTime="2026-03-14 09:51:43.056533628 +0000 UTC m=+2470.369144756" Mar 14 09:51:43 crc kubenswrapper[4843]: I0314 09:51:43.067536 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.067516867 podStartE2EDuration="2.067516867s" podCreationTimestamp="2026-03-14 09:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:51:43.067339563 +0000 UTC m=+2470.379950691" watchObservedRunningTime="2026-03-14 09:51:43.067516867 +0000 UTC m=+2470.380127995" Mar 14 09:51:45 crc kubenswrapper[4843]: I0314 09:51:45.034084 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:51:45 crc kubenswrapper[4843]: I0314 09:51:45.507470 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:46 crc kubenswrapper[4843]: I0314 09:51:46.573904 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:46 crc kubenswrapper[4843]: I0314 09:51:46.590427 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:47 crc kubenswrapper[4843]: I0314 09:51:47.339383 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:51:47 crc kubenswrapper[4843]: E0314 09:51:47.339695 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:51:51 crc kubenswrapper[4843]: I0314 09:51:51.575095 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:51 crc kubenswrapper[4843]: I0314 09:51:51.581899 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:51 crc kubenswrapper[4843]: I0314 09:51:51.590418 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:51 crc kubenswrapper[4843]: I0314 09:51:51.624070 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:51 crc kubenswrapper[4843]: I0314 09:51:51.764897 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:51 crc kubenswrapper[4843]: I0314 09:51:51.788315 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:52 crc kubenswrapper[4843]: I0314 09:51:52.087352 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:52 crc kubenswrapper[4843]: I0314 09:51:52.098595 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:51:52 crc kubenswrapper[4843]: I0314 09:51:52.134964 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:51:52 crc kubenswrapper[4843]: I0314 09:51:52.138595 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:51:54 crc kubenswrapper[4843]: I0314 09:51:54.387669 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:54 crc kubenswrapper[4843]: I0314 09:51:54.388315 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-central-agent" containerID="cri-o://cda2091bf4f3e6e0e5a74f0d2259fdde47b20124b53fd77d8fa80662968fddbc" gracePeriod=30 Mar 14 09:51:54 crc kubenswrapper[4843]: I0314 09:51:54.388371 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="sg-core" containerID="cri-o://3cbd0922fb7c2a426e4a6f1b87cfca5ed741f1b809766b207df4950d48efdacf" gracePeriod=30 Mar 14 09:51:54 crc kubenswrapper[4843]: I0314 09:51:54.388381 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="proxy-httpd" containerID="cri-o://419165322e603cd53c2137dbb1d5630a01f35bf58e5bb16bb782c5da7958987e" gracePeriod=30 Mar 14 09:51:54 crc kubenswrapper[4843]: I0314 09:51:54.388418 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-notification-agent" containerID="cri-o://e478bfbbf9b07eb8b1c597b506f27d645773dd7382092926e5d126c7cd23b643" gracePeriod=30 Mar 14 09:51:54 crc kubenswrapper[4843]: I0314 09:51:54.495860 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.207:3000/\": read tcp 10.217.0.2:44708->10.217.0.207:3000: read: connection reset by peer" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.128860 4843 generic.go:334] "Generic (PLEG): container finished" podID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerID="419165322e603cd53c2137dbb1d5630a01f35bf58e5bb16bb782c5da7958987e" exitCode=0 Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.129546 4843 generic.go:334] "Generic (PLEG): container finished" podID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerID="3cbd0922fb7c2a426e4a6f1b87cfca5ed741f1b809766b207df4950d48efdacf" exitCode=2 Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.129562 4843 generic.go:334] "Generic (PLEG): container finished" podID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerID="e478bfbbf9b07eb8b1c597b506f27d645773dd7382092926e5d126c7cd23b643" exitCode=0 Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.129569 4843 generic.go:334] "Generic (PLEG): container finished" podID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerID="cda2091bf4f3e6e0e5a74f0d2259fdde47b20124b53fd77d8fa80662968fddbc" exitCode=0 Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.128916 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerDied","Data":"419165322e603cd53c2137dbb1d5630a01f35bf58e5bb16bb782c5da7958987e"} Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.129604 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerDied","Data":"3cbd0922fb7c2a426e4a6f1b87cfca5ed741f1b809766b207df4950d48efdacf"} Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.129699 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerDied","Data":"e478bfbbf9b07eb8b1c597b506f27d645773dd7382092926e5d126c7cd23b643"} Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.129712 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerDied","Data":"cda2091bf4f3e6e0e5a74f0d2259fdde47b20124b53fd77d8fa80662968fddbc"} Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.199307 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.310931 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-config-data\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311051 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n67np\" (UniqueName: \"kubernetes.io/projected/441d7f37-0bad-4a93-94f6-71d380ca73b1-kube-api-access-n67np\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311109 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-run-httpd\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311156 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-ceilometer-tls-certs\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311206 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-combined-ca-bundle\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311240 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-scripts\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311292 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-sg-core-conf-yaml\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311385 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-log-httpd\") pod \"441d7f37-0bad-4a93-94f6-71d380ca73b1\" (UID: \"441d7f37-0bad-4a93-94f6-71d380ca73b1\") " Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.311537 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.312086 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.312203 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.315987 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/441d7f37-0bad-4a93-94f6-71d380ca73b1-kube-api-access-n67np" (OuterVolumeSpecName: "kube-api-access-n67np") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "kube-api-access-n67np". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.316357 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-scripts" (OuterVolumeSpecName: "scripts") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.350901 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.361947 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.382764 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.390101 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-config-data" (OuterVolumeSpecName: "config-data") pod "441d7f37-0bad-4a93-94f6-71d380ca73b1" (UID: "441d7f37-0bad-4a93-94f6-71d380ca73b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413577 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441d7f37-0bad-4a93-94f6-71d380ca73b1-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413613 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413623 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n67np\" (UniqueName: \"kubernetes.io/projected/441d7f37-0bad-4a93-94f6-71d380ca73b1-kube-api-access-n67np\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413633 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413642 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413651 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:55 crc kubenswrapper[4843]: I0314 09:51:55.413660 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441d7f37-0bad-4a93-94f6-71d380ca73b1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.149187 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"441d7f37-0bad-4a93-94f6-71d380ca73b1","Type":"ContainerDied","Data":"2360222e37d4cec9d7419bf07885505b14c52ed2e032352b086885b94d94deb4"} Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.149534 4843 scope.go:117] "RemoveContainer" containerID="419165322e603cd53c2137dbb1d5630a01f35bf58e5bb16bb782c5da7958987e" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.149686 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.186591 4843 scope.go:117] "RemoveContainer" containerID="3cbd0922fb7c2a426e4a6f1b87cfca5ed741f1b809766b207df4950d48efdacf" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.192861 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.221426 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.227158 4843 scope.go:117] "RemoveContainer" containerID="e478bfbbf9b07eb8b1c597b506f27d645773dd7382092926e5d126c7cd23b643" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.246542 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:56 crc kubenswrapper[4843]: E0314 09:51:56.247073 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="sg-core" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247094 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="sg-core" Mar 14 09:51:56 crc kubenswrapper[4843]: E0314 09:51:56.247103 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-central-agent" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247110 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-central-agent" Mar 14 09:51:56 crc kubenswrapper[4843]: E0314 09:51:56.247124 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-notification-agent" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247132 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-notification-agent" Mar 14 09:51:56 crc kubenswrapper[4843]: E0314 09:51:56.247144 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="proxy-httpd" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247151 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="proxy-httpd" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247380 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="proxy-httpd" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247392 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-central-agent" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247406 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="sg-core" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.247416 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" containerName="ceilometer-notification-agent" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.249261 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.253065 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.261081 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.261403 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.261588 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.266985 4843 scope.go:117] "RemoveContainer" containerID="cda2091bf4f3e6e0e5a74f0d2259fdde47b20124b53fd77d8fa80662968fddbc" Mar 14 09:51:56 crc kubenswrapper[4843]: E0314 09:51:56.346619 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod441d7f37_0bad_4a93_94f6_71d380ca73b1.slice/crio-2360222e37d4cec9d7419bf07885505b14c52ed2e032352b086885b94d94deb4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod441d7f37_0bad_4a93_94f6_71d380ca73b1.slice\": RecentStats: unable to find data in memory cache]" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.434838 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.434892 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.435018 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-config-data\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.435082 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-run-httpd\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.435132 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.435160 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-log-httpd\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.435189 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlt4w\" (UniqueName: \"kubernetes.io/projected/aebd0ba2-ff58-46a4-9ecf-1ada02952691-kube-api-access-vlt4w\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.435219 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-scripts\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536360 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-config-data\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536403 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-run-httpd\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536435 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536455 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-log-httpd\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536476 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlt4w\" (UniqueName: \"kubernetes.io/projected/aebd0ba2-ff58-46a4-9ecf-1ada02952691-kube-api-access-vlt4w\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536496 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-scripts\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536536 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.536551 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.537503 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-log-httpd\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.537531 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-run-httpd\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.543115 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.543682 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.543922 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-scripts\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.544238 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-config-data\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.553432 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.554092 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlt4w\" (UniqueName: \"kubernetes.io/projected/aebd0ba2-ff58-46a4-9ecf-1ada02952691-kube-api-access-vlt4w\") pod \"ceilometer-0\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:56 crc kubenswrapper[4843]: I0314 09:51:56.575415 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:51:57 crc kubenswrapper[4843]: I0314 09:51:57.034185 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:51:57 crc kubenswrapper[4843]: I0314 09:51:57.162548 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerStarted","Data":"11b7c1253359ed3ddd4b6992146708fc983f9d5ec35bdd3c789c16a0d53240e5"} Mar 14 09:51:57 crc kubenswrapper[4843]: I0314 09:51:57.352015 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="441d7f37-0bad-4a93-94f6-71d380ca73b1" path="/var/lib/kubelet/pods/441d7f37-0bad-4a93-94f6-71d380ca73b1/volumes" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.171828 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerStarted","Data":"d49b3a8d5a4b2cc65bf1517026b7db84fc6a94adb636d6f5b5c08d0e9778e83b"} Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.187651 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.193853 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-sk4w8"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.256468 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.256687 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="9b226653-0262-4cdb-9e8a-d3681331d87c" containerName="watcher-applier" containerID="cri-o://2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" gracePeriod=30 Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.324576 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.324795 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-kuttl-api-log" containerID="cri-o://ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf" gracePeriod=30 Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.325144 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-api" containerID="cri-o://5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e" gracePeriod=30 Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.359572 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.359766 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="bba9218c-c199-4a70-bd95-f1c1de972639" containerName="watcher-decision-engine" containerID="cri-o://1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74" gracePeriod=30 Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.652068 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher5b3a-account-delete-5jvnq"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.657500 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.668341 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher5b3a-account-delete-5jvnq"] Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.784516 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65mr8\" (UniqueName: \"kubernetes.io/projected/b55d408e-d1c2-4fe1-82b8-05e72c669dba-kube-api-access-65mr8\") pod \"watcher5b3a-account-delete-5jvnq\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.784792 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b55d408e-d1c2-4fe1-82b8-05e72c669dba-operator-scripts\") pod \"watcher5b3a-account-delete-5jvnq\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.887357 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65mr8\" (UniqueName: \"kubernetes.io/projected/b55d408e-d1c2-4fe1-82b8-05e72c669dba-kube-api-access-65mr8\") pod \"watcher5b3a-account-delete-5jvnq\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.887420 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b55d408e-d1c2-4fe1-82b8-05e72c669dba-operator-scripts\") pod \"watcher5b3a-account-delete-5jvnq\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.888418 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b55d408e-d1c2-4fe1-82b8-05e72c669dba-operator-scripts\") pod \"watcher5b3a-account-delete-5jvnq\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:58 crc kubenswrapper[4843]: I0314 09:51:58.909705 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65mr8\" (UniqueName: \"kubernetes.io/projected/b55d408e-d1c2-4fe1-82b8-05e72c669dba-kube-api-access-65mr8\") pod \"watcher5b3a-account-delete-5jvnq\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:59 crc kubenswrapper[4843]: I0314 09:51:59.006424 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:51:59 crc kubenswrapper[4843]: I0314 09:51:59.227648 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerStarted","Data":"1d192e24e961d679f3dd74c7d486a265163f773a9ec5ff03d1aa62fe4ce5c31e"} Mar 14 09:51:59 crc kubenswrapper[4843]: I0314 09:51:59.229556 4843 generic.go:334] "Generic (PLEG): container finished" podID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerID="ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf" exitCode=143 Mar 14 09:51:59 crc kubenswrapper[4843]: I0314 09:51:59.229596 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"f73f2e72-f036-411d-9e88-c52f5795ec97","Type":"ContainerDied","Data":"ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf"} Mar 14 09:51:59 crc kubenswrapper[4843]: I0314 09:51:59.348809 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f097b765-f4be-4e04-a4c9-8391e58aaf20" path="/var/lib/kubelet/pods/f097b765-f4be-4e04-a4c9-8391e58aaf20/volumes" Mar 14 09:51:59 crc kubenswrapper[4843]: I0314 09:51:59.548544 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher5b3a-account-delete-5jvnq"] Mar 14 09:51:59 crc kubenswrapper[4843]: W0314 09:51:59.556974 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb55d408e_d1c2_4fe1_82b8_05e72c669dba.slice/crio-260e6fe1a9c67c3c07b8bb12f6e3d541b2a04b3a6a7612d4bbabd064aeb37ab3 WatchSource:0}: Error finding container 260e6fe1a9c67c3c07b8bb12f6e3d541b2a04b3a6a7612d4bbabd064aeb37ab3: Status 404 returned error can't find the container with id 260e6fe1a9c67c3c07b8bb12f6e3d541b2a04b3a6a7612d4bbabd064aeb37ab3 Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.138314 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558032-ms2dx"] Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.139769 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.142181 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.143080 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.143222 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.145461 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558032-ms2dx"] Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.214386 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f5vq\" (UniqueName: \"kubernetes.io/projected/33f82e98-788c-4584-bc61-3ed2704f00df-kube-api-access-6f5vq\") pod \"auto-csr-approver-29558032-ms2dx\" (UID: \"33f82e98-788c-4584-bc61-3ed2704f00df\") " pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.215012 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.237915 4843 generic.go:334] "Generic (PLEG): container finished" podID="b55d408e-d1c2-4fe1-82b8-05e72c669dba" containerID="11061fdb195c33c214175c1687a71b05388167237fb3c60cb085c9a78c6ca148" exitCode=0 Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.238106 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" event={"ID":"b55d408e-d1c2-4fe1-82b8-05e72c669dba","Type":"ContainerDied","Data":"11061fdb195c33c214175c1687a71b05388167237fb3c60cb085c9a78c6ca148"} Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.238138 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" event={"ID":"b55d408e-d1c2-4fe1-82b8-05e72c669dba","Type":"ContainerStarted","Data":"260e6fe1a9c67c3c07b8bb12f6e3d541b2a04b3a6a7612d4bbabd064aeb37ab3"} Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.241001 4843 generic.go:334] "Generic (PLEG): container finished" podID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerID="5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e" exitCode=0 Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.241054 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"f73f2e72-f036-411d-9e88-c52f5795ec97","Type":"ContainerDied","Data":"5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e"} Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.241077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"f73f2e72-f036-411d-9e88-c52f5795ec97","Type":"ContainerDied","Data":"6371eeeade059315881a6873ef08f350a585471674b22de7d7ab289ccbc2e7f0"} Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.241080 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.241093 4843 scope.go:117] "RemoveContainer" containerID="5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.247021 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerStarted","Data":"5819a34913f0ac2503fdac74948d9e4f0dea738e203835018f6c2e55df34d9e2"} Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.261158 4843 scope.go:117] "RemoveContainer" containerID="ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.280605 4843 scope.go:117] "RemoveContainer" containerID="5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e" Mar 14 09:52:00 crc kubenswrapper[4843]: E0314 09:52:00.281244 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e\": container with ID starting with 5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e not found: ID does not exist" containerID="5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.281322 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e"} err="failed to get container status \"5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e\": rpc error: code = NotFound desc = could not find container \"5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e\": container with ID starting with 5b8aa5acbf908a3c06d698a57400de093ccaf3414fb3c5ee61aba7213055fb3e not found: ID does not exist" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.281356 4843 scope.go:117] "RemoveContainer" containerID="ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf" Mar 14 09:52:00 crc kubenswrapper[4843]: E0314 09:52:00.281759 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf\": container with ID starting with ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf not found: ID does not exist" containerID="ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.281806 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf"} err="failed to get container status \"ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf\": rpc error: code = NotFound desc = could not find container \"ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf\": container with ID starting with ffccd874e604fa432276a354a2d7be66b1891ab6996e309167556e9ccd17c5bf not found: ID does not exist" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.316522 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-custom-prometheus-ca\") pod \"f73f2e72-f036-411d-9e88-c52f5795ec97\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.316619 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f73f2e72-f036-411d-9e88-c52f5795ec97-logs\") pod \"f73f2e72-f036-411d-9e88-c52f5795ec97\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.316724 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-cert-memcached-mtls\") pod \"f73f2e72-f036-411d-9e88-c52f5795ec97\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.316802 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-combined-ca-bundle\") pod \"f73f2e72-f036-411d-9e88-c52f5795ec97\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.316849 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbftp\" (UniqueName: \"kubernetes.io/projected/f73f2e72-f036-411d-9e88-c52f5795ec97-kube-api-access-zbftp\") pod \"f73f2e72-f036-411d-9e88-c52f5795ec97\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.316901 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-config-data\") pod \"f73f2e72-f036-411d-9e88-c52f5795ec97\" (UID: \"f73f2e72-f036-411d-9e88-c52f5795ec97\") " Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.317186 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f5vq\" (UniqueName: \"kubernetes.io/projected/33f82e98-788c-4584-bc61-3ed2704f00df-kube-api-access-6f5vq\") pod \"auto-csr-approver-29558032-ms2dx\" (UID: \"33f82e98-788c-4584-bc61-3ed2704f00df\") " pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.317468 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f73f2e72-f036-411d-9e88-c52f5795ec97-logs" (OuterVolumeSpecName: "logs") pod "f73f2e72-f036-411d-9e88-c52f5795ec97" (UID: "f73f2e72-f036-411d-9e88-c52f5795ec97"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.332360 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f73f2e72-f036-411d-9e88-c52f5795ec97-kube-api-access-zbftp" (OuterVolumeSpecName: "kube-api-access-zbftp") pod "f73f2e72-f036-411d-9e88-c52f5795ec97" (UID: "f73f2e72-f036-411d-9e88-c52f5795ec97"). InnerVolumeSpecName "kube-api-access-zbftp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.332759 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f5vq\" (UniqueName: \"kubernetes.io/projected/33f82e98-788c-4584-bc61-3ed2704f00df-kube-api-access-6f5vq\") pod \"auto-csr-approver-29558032-ms2dx\" (UID: \"33f82e98-788c-4584-bc61-3ed2704f00df\") " pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.349413 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f73f2e72-f036-411d-9e88-c52f5795ec97" (UID: "f73f2e72-f036-411d-9e88-c52f5795ec97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.349913 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "f73f2e72-f036-411d-9e88-c52f5795ec97" (UID: "f73f2e72-f036-411d-9e88-c52f5795ec97"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.388521 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "f73f2e72-f036-411d-9e88-c52f5795ec97" (UID: "f73f2e72-f036-411d-9e88-c52f5795ec97"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.390997 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-config-data" (OuterVolumeSpecName: "config-data") pod "f73f2e72-f036-411d-9e88-c52f5795ec97" (UID: "f73f2e72-f036-411d-9e88-c52f5795ec97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.418823 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.418868 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.418883 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f73f2e72-f036-411d-9e88-c52f5795ec97-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.418895 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.418909 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73f2e72-f036-411d-9e88-c52f5795ec97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.418921 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbftp\" (UniqueName: \"kubernetes.io/projected/f73f2e72-f036-411d-9e88-c52f5795ec97-kube-api-access-zbftp\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.524397 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.581355 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.613000 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:00 crc kubenswrapper[4843]: I0314 09:52:00.987542 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558032-ms2dx"] Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.256879 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" event={"ID":"33f82e98-788c-4584-bc61-3ed2704f00df","Type":"ContainerStarted","Data":"76028c178a7debbbfaa0d9e785a3b18444e5152052661fd6b81f246ab32a7473"} Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.361055 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" path="/var/lib/kubelet/pods/f73f2e72-f036-411d-9e88-c52f5795ec97/volumes" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.517829 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:01 crc kubenswrapper[4843]: E0314 09:52:01.594229 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b is running failed: container process not found" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:52:01 crc kubenswrapper[4843]: E0314 09:52:01.600545 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b is running failed: container process not found" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:52:01 crc kubenswrapper[4843]: E0314 09:52:01.609420 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b is running failed: container process not found" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:52:01 crc kubenswrapper[4843]: E0314 09:52:01.609496 4843 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b is running failed: container process not found" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="9b226653-0262-4cdb-9e8a-d3681331d87c" containerName="watcher-applier" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.609631 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.665089 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.741431 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b55d408e-d1c2-4fe1-82b8-05e72c669dba-operator-scripts\") pod \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.741740 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65mr8\" (UniqueName: \"kubernetes.io/projected/b55d408e-d1c2-4fe1-82b8-05e72c669dba-kube-api-access-65mr8\") pod \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\" (UID: \"b55d408e-d1c2-4fe1-82b8-05e72c669dba\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.746513 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b55d408e-d1c2-4fe1-82b8-05e72c669dba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b55d408e-d1c2-4fe1-82b8-05e72c669dba" (UID: "b55d408e-d1c2-4fe1-82b8-05e72c669dba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.746907 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b55d408e-d1c2-4fe1-82b8-05e72c669dba-kube-api-access-65mr8" (OuterVolumeSpecName: "kube-api-access-65mr8") pod "b55d408e-d1c2-4fe1-82b8-05e72c669dba" (UID: "b55d408e-d1c2-4fe1-82b8-05e72c669dba"). InnerVolumeSpecName "kube-api-access-65mr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.843569 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-combined-ca-bundle\") pod \"9b226653-0262-4cdb-9e8a-d3681331d87c\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.843641 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-cert-memcached-mtls\") pod \"9b226653-0262-4cdb-9e8a-d3681331d87c\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.843813 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwq28\" (UniqueName: \"kubernetes.io/projected/9b226653-0262-4cdb-9e8a-d3681331d87c-kube-api-access-gwq28\") pod \"9b226653-0262-4cdb-9e8a-d3681331d87c\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.844006 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-config-data\") pod \"9b226653-0262-4cdb-9e8a-d3681331d87c\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.844056 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b226653-0262-4cdb-9e8a-d3681331d87c-logs\") pod \"9b226653-0262-4cdb-9e8a-d3681331d87c\" (UID: \"9b226653-0262-4cdb-9e8a-d3681331d87c\") " Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.844511 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b226653-0262-4cdb-9e8a-d3681331d87c-logs" (OuterVolumeSpecName: "logs") pod "9b226653-0262-4cdb-9e8a-d3681331d87c" (UID: "9b226653-0262-4cdb-9e8a-d3681331d87c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.844792 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b55d408e-d1c2-4fe1-82b8-05e72c669dba-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.844818 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65mr8\" (UniqueName: \"kubernetes.io/projected/b55d408e-d1c2-4fe1-82b8-05e72c669dba-kube-api-access-65mr8\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.844835 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b226653-0262-4cdb-9e8a-d3681331d87c-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.848622 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b226653-0262-4cdb-9e8a-d3681331d87c-kube-api-access-gwq28" (OuterVolumeSpecName: "kube-api-access-gwq28") pod "9b226653-0262-4cdb-9e8a-d3681331d87c" (UID: "9b226653-0262-4cdb-9e8a-d3681331d87c"). InnerVolumeSpecName "kube-api-access-gwq28". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.873598 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b226653-0262-4cdb-9e8a-d3681331d87c" (UID: "9b226653-0262-4cdb-9e8a-d3681331d87c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.905440 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-config-data" (OuterVolumeSpecName: "config-data") pod "9b226653-0262-4cdb-9e8a-d3681331d87c" (UID: "9b226653-0262-4cdb-9e8a-d3681331d87c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.946713 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.946757 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.946771 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwq28\" (UniqueName: \"kubernetes.io/projected/9b226653-0262-4cdb-9e8a-d3681331d87c-kube-api-access-gwq28\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:01 crc kubenswrapper[4843]: I0314 09:52:01.958429 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "9b226653-0262-4cdb-9e8a-d3681331d87c" (UID: "9b226653-0262-4cdb-9e8a-d3681331d87c"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.048938 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9b226653-0262-4cdb-9e8a-d3681331d87c-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.272116 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.272341 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher5b3a-account-delete-5jvnq" event={"ID":"b55d408e-d1c2-4fe1-82b8-05e72c669dba","Type":"ContainerDied","Data":"260e6fe1a9c67c3c07b8bb12f6e3d541b2a04b3a6a7612d4bbabd064aeb37ab3"} Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.272567 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="260e6fe1a9c67c3c07b8bb12f6e3d541b2a04b3a6a7612d4bbabd064aeb37ab3" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.274948 4843 generic.go:334] "Generic (PLEG): container finished" podID="9b226653-0262-4cdb-9e8a-d3681331d87c" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" exitCode=0 Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.275042 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"9b226653-0262-4cdb-9e8a-d3681331d87c","Type":"ContainerDied","Data":"2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b"} Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.275088 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"9b226653-0262-4cdb-9e8a-d3681331d87c","Type":"ContainerDied","Data":"cc9244c998471e557a992640588c115509e31a3effeb76c0f25ee6aa4c710960"} Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.275105 4843 scope.go:117] "RemoveContainer" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.275248 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.278571 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerStarted","Data":"044ef9eaaf6c0241839a838b71f7d65c292a92a4ff23baa5f8d469e9cb3c0e35"} Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.278704 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-central-agent" containerID="cri-o://d49b3a8d5a4b2cc65bf1517026b7db84fc6a94adb636d6f5b5c08d0e9778e83b" gracePeriod=30 Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.278891 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.278933 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="proxy-httpd" containerID="cri-o://044ef9eaaf6c0241839a838b71f7d65c292a92a4ff23baa5f8d469e9cb3c0e35" gracePeriod=30 Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.278975 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="sg-core" containerID="cri-o://5819a34913f0ac2503fdac74948d9e4f0dea738e203835018f6c2e55df34d9e2" gracePeriod=30 Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.279014 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-notification-agent" containerID="cri-o://1d192e24e961d679f3dd74c7d486a265163f773a9ec5ff03d1aa62fe4ce5c31e" gracePeriod=30 Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.300045 4843 scope.go:117] "RemoveContainer" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" Mar 14 09:52:02 crc kubenswrapper[4843]: E0314 09:52:02.302458 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b\": container with ID starting with 2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b not found: ID does not exist" containerID="2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.302527 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b"} err="failed to get container status \"2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b\": rpc error: code = NotFound desc = could not find container \"2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b\": container with ID starting with 2b4fc5b20259365f574d39697364a7f4622aa8a63957d7121da32ced4eeef36b not found: ID does not exist" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.306657 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.864714884 podStartE2EDuration="6.306635037s" podCreationTimestamp="2026-03-14 09:51:56 +0000 UTC" firstStartedPulling="2026-03-14 09:51:57.040229624 +0000 UTC m=+2484.352840762" lastFinishedPulling="2026-03-14 09:52:01.482149777 +0000 UTC m=+2488.794760915" observedRunningTime="2026-03-14 09:52:02.299245185 +0000 UTC m=+2489.611856323" watchObservedRunningTime="2026-03-14 09:52:02.306635037 +0000 UTC m=+2489.619246165" Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.326143 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.333636 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:02 crc kubenswrapper[4843]: I0314 09:52:02.339831 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:52:02 crc kubenswrapper[4843]: E0314 09:52:02.340017 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.061579 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.166559 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-custom-prometheus-ca\") pod \"bba9218c-c199-4a70-bd95-f1c1de972639\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.166686 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-combined-ca-bundle\") pod \"bba9218c-c199-4a70-bd95-f1c1de972639\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.166731 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkjnf\" (UniqueName: \"kubernetes.io/projected/bba9218c-c199-4a70-bd95-f1c1de972639-kube-api-access-rkjnf\") pod \"bba9218c-c199-4a70-bd95-f1c1de972639\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.166754 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-cert-memcached-mtls\") pod \"bba9218c-c199-4a70-bd95-f1c1de972639\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.166783 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bba9218c-c199-4a70-bd95-f1c1de972639-logs\") pod \"bba9218c-c199-4a70-bd95-f1c1de972639\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.166846 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-config-data\") pod \"bba9218c-c199-4a70-bd95-f1c1de972639\" (UID: \"bba9218c-c199-4a70-bd95-f1c1de972639\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.170407 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba9218c-c199-4a70-bd95-f1c1de972639-logs" (OuterVolumeSpecName: "logs") pod "bba9218c-c199-4a70-bd95-f1c1de972639" (UID: "bba9218c-c199-4a70-bd95-f1c1de972639"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.185882 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bba9218c-c199-4a70-bd95-f1c1de972639-kube-api-access-rkjnf" (OuterVolumeSpecName: "kube-api-access-rkjnf") pod "bba9218c-c199-4a70-bd95-f1c1de972639" (UID: "bba9218c-c199-4a70-bd95-f1c1de972639"). InnerVolumeSpecName "kube-api-access-rkjnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.190862 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bba9218c-c199-4a70-bd95-f1c1de972639" (UID: "bba9218c-c199-4a70-bd95-f1c1de972639"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.203858 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "bba9218c-c199-4a70-bd95-f1c1de972639" (UID: "bba9218c-c199-4a70-bd95-f1c1de972639"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.216469 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-config-data" (OuterVolumeSpecName: "config-data") pod "bba9218c-c199-4a70-bd95-f1c1de972639" (UID: "bba9218c-c199-4a70-bd95-f1c1de972639"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.229090 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "bba9218c-c199-4a70-bd95-f1c1de972639" (UID: "bba9218c-c199-4a70-bd95-f1c1de972639"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.269313 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bba9218c-c199-4a70-bd95-f1c1de972639-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.269349 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.269360 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.269371 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.269384 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkjnf\" (UniqueName: \"kubernetes.io/projected/bba9218c-c199-4a70-bd95-f1c1de972639-kube-api-access-rkjnf\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.269392 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/bba9218c-c199-4a70-bd95-f1c1de972639-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.297326 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-nx2zd"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.309998 4843 generic.go:334] "Generic (PLEG): container finished" podID="bba9218c-c199-4a70-bd95-f1c1de972639" containerID="1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74" exitCode=0 Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.310056 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"bba9218c-c199-4a70-bd95-f1c1de972639","Type":"ContainerDied","Data":"1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.310080 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"bba9218c-c199-4a70-bd95-f1c1de972639","Type":"ContainerDied","Data":"8654108872b95a523010e371ccc79e2717f140ef7d8f597ee8d706afcaf019bc"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.310095 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-nx2zd"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.310127 4843 scope.go:117] "RemoveContainer" containerID="1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.310210 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.318707 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319375 4843 generic.go:334] "Generic (PLEG): container finished" podID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerID="044ef9eaaf6c0241839a838b71f7d65c292a92a4ff23baa5f8d469e9cb3c0e35" exitCode=0 Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319399 4843 generic.go:334] "Generic (PLEG): container finished" podID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerID="5819a34913f0ac2503fdac74948d9e4f0dea738e203835018f6c2e55df34d9e2" exitCode=2 Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319406 4843 generic.go:334] "Generic (PLEG): container finished" podID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerID="1d192e24e961d679f3dd74c7d486a265163f773a9ec5ff03d1aa62fe4ce5c31e" exitCode=0 Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319413 4843 generic.go:334] "Generic (PLEG): container finished" podID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerID="d49b3a8d5a4b2cc65bf1517026b7db84fc6a94adb636d6f5b5c08d0e9778e83b" exitCode=0 Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319420 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerDied","Data":"044ef9eaaf6c0241839a838b71f7d65c292a92a4ff23baa5f8d469e9cb3c0e35"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319451 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerDied","Data":"5819a34913f0ac2503fdac74948d9e4f0dea738e203835018f6c2e55df34d9e2"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319467 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerDied","Data":"1d192e24e961d679f3dd74c7d486a265163f773a9ec5ff03d1aa62fe4ce5c31e"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.319486 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerDied","Data":"d49b3a8d5a4b2cc65bf1517026b7db84fc6a94adb636d6f5b5c08d0e9778e83b"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.320890 4843 generic.go:334] "Generic (PLEG): container finished" podID="33f82e98-788c-4584-bc61-3ed2704f00df" containerID="59e7029d842464cc8f2d184fae3003c0c2d520b17db970dc9b61f9f6949f03f2" exitCode=0 Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.320918 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" event={"ID":"33f82e98-788c-4584-bc61-3ed2704f00df","Type":"ContainerDied","Data":"59e7029d842464cc8f2d184fae3003c0c2d520b17db970dc9b61f9f6949f03f2"} Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.326435 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher5b3a-account-delete-5jvnq"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.336221 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-5b3a-account-create-update-rlchp"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.343103 4843 scope.go:117] "RemoveContainer" containerID="1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74" Mar 14 09:52:03 crc kubenswrapper[4843]: E0314 09:52:03.343499 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74\": container with ID starting with 1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74 not found: ID does not exist" containerID="1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.343538 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74"} err="failed to get container status \"1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74\": rpc error: code = NotFound desc = could not find container \"1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74\": container with ID starting with 1146fbe432716101281bd9a6dfcb0fb7b996c521e3cf4c3591f02abb1ec21b74 not found: ID does not exist" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.358036 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd8d7c1-b35b-4982-9cf4-04ed8a29192f" path="/var/lib/kubelet/pods/1fd8d7c1-b35b-4982-9cf4-04ed8a29192f/volumes" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.358565 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="959c17a9-c974-49e8-bf5a-60be6bf6a944" path="/var/lib/kubelet/pods/959c17a9-c974-49e8-bf5a-60be6bf6a944/volumes" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.359060 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b226653-0262-4cdb-9e8a-d3681331d87c" path="/var/lib/kubelet/pods/9b226653-0262-4cdb-9e8a-d3681331d87c/volumes" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.359598 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher5b3a-account-delete-5jvnq"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.364241 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.369103 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.371476 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471369 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-run-httpd\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471468 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-sg-core-conf-yaml\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471504 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-scripts\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471544 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-log-httpd\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471563 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-config-data\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471609 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-combined-ca-bundle\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471644 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-ceilometer-tls-certs\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.471729 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlt4w\" (UniqueName: \"kubernetes.io/projected/aebd0ba2-ff58-46a4-9ecf-1ada02952691-kube-api-access-vlt4w\") pod \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\" (UID: \"aebd0ba2-ff58-46a4-9ecf-1ada02952691\") " Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.473241 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.474588 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.480211 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-scripts" (OuterVolumeSpecName: "scripts") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.480293 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aebd0ba2-ff58-46a4-9ecf-1ada02952691-kube-api-access-vlt4w" (OuterVolumeSpecName: "kube-api-access-vlt4w") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "kube-api-access-vlt4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.493584 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.509963 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.551529 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-config-data" (OuterVolumeSpecName: "config-data") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.558312 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aebd0ba2-ff58-46a4-9ecf-1ada02952691" (UID: "aebd0ba2-ff58-46a4-9ecf-1ada02952691"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573388 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573430 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573441 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573451 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573459 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aebd0ba2-ff58-46a4-9ecf-1ada02952691-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573467 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573476 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aebd0ba2-ff58-46a4-9ecf-1ada02952691-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:03 crc kubenswrapper[4843]: I0314 09:52:03.573486 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlt4w\" (UniqueName: \"kubernetes.io/projected/aebd0ba2-ff58-46a4-9ecf-1ada02952691-kube-api-access-vlt4w\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.263515 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-fzhnw"] Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.263932 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba9218c-c199-4a70-bd95-f1c1de972639" containerName="watcher-decision-engine" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.263959 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba9218c-c199-4a70-bd95-f1c1de972639" containerName="watcher-decision-engine" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.263969 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55d408e-d1c2-4fe1-82b8-05e72c669dba" containerName="mariadb-account-delete" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.263976 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55d408e-d1c2-4fe1-82b8-05e72c669dba" containerName="mariadb-account-delete" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.263987 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="proxy-httpd" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264013 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="proxy-httpd" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.264031 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-kuttl-api-log" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264039 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-kuttl-api-log" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.264057 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-notification-agent" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264065 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-notification-agent" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.264079 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="sg-core" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264086 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="sg-core" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.264097 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-api" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264104 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-api" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.264124 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b226653-0262-4cdb-9e8a-d3681331d87c" containerName="watcher-applier" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264132 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b226653-0262-4cdb-9e8a-d3681331d87c" containerName="watcher-applier" Mar 14 09:52:04 crc kubenswrapper[4843]: E0314 09:52:04.264150 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-central-agent" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264158 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-central-agent" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264349 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="proxy-httpd" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264371 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-central-agent" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264382 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bba9218c-c199-4a70-bd95-f1c1de972639" containerName="watcher-decision-engine" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264393 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="ceilometer-notification-agent" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264409 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b55d408e-d1c2-4fe1-82b8-05e72c669dba" containerName="mariadb-account-delete" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264419 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b226653-0262-4cdb-9e8a-d3681331d87c" containerName="watcher-applier" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264428 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-kuttl-api-log" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264441 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" containerName="sg-core" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.264453 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73f2e72-f036-411d-9e88-c52f5795ec97" containerName="watcher-api" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.265140 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.276434 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-25dc-account-create-update-qssln"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.277374 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-fzhnw"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.277454 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.312200 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef4afd5-c0a9-4443-aa51-358adb5b9014-operator-scripts\") pod \"watcher-25dc-account-create-update-qssln\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.312287 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqsmx\" (UniqueName: \"kubernetes.io/projected/bef4afd5-c0a9-4443-aa51-358adb5b9014-kube-api-access-pqsmx\") pod \"watcher-25dc-account-create-update-qssln\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.312330 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-operator-scripts\") pod \"watcher-db-create-fzhnw\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.312558 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cclg\" (UniqueName: \"kubernetes.io/projected/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-kube-api-access-6cclg\") pod \"watcher-db-create-fzhnw\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.318296 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.342361 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.342460 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"aebd0ba2-ff58-46a4-9ecf-1ada02952691","Type":"ContainerDied","Data":"11b7c1253359ed3ddd4b6992146708fc983f9d5ec35bdd3c789c16a0d53240e5"} Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.342633 4843 scope.go:117] "RemoveContainer" containerID="044ef9eaaf6c0241839a838b71f7d65c292a92a4ff23baa5f8d469e9cb3c0e35" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.344155 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-25dc-account-create-update-qssln"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.376692 4843 scope.go:117] "RemoveContainer" containerID="5819a34913f0ac2503fdac74948d9e4f0dea738e203835018f6c2e55df34d9e2" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.398330 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.412550 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.414124 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef4afd5-c0a9-4443-aa51-358adb5b9014-operator-scripts\") pod \"watcher-25dc-account-create-update-qssln\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.414385 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqsmx\" (UniqueName: \"kubernetes.io/projected/bef4afd5-c0a9-4443-aa51-358adb5b9014-kube-api-access-pqsmx\") pod \"watcher-25dc-account-create-update-qssln\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.414478 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-operator-scripts\") pod \"watcher-db-create-fzhnw\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.414573 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cclg\" (UniqueName: \"kubernetes.io/projected/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-kube-api-access-6cclg\") pod \"watcher-db-create-fzhnw\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.415442 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-operator-scripts\") pod \"watcher-db-create-fzhnw\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.415703 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef4afd5-c0a9-4443-aa51-358adb5b9014-operator-scripts\") pod \"watcher-25dc-account-create-update-qssln\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.428058 4843 scope.go:117] "RemoveContainer" containerID="1d192e24e961d679f3dd74c7d486a265163f773a9ec5ff03d1aa62fe4ce5c31e" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.433337 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.436076 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.438952 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.439157 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.439378 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.445146 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqsmx\" (UniqueName: \"kubernetes.io/projected/bef4afd5-c0a9-4443-aa51-358adb5b9014-kube-api-access-pqsmx\") pod \"watcher-25dc-account-create-update-qssln\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.451290 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.461771 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cclg\" (UniqueName: \"kubernetes.io/projected/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-kube-api-access-6cclg\") pod \"watcher-db-create-fzhnw\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.475522 4843 scope.go:117] "RemoveContainer" containerID="d49b3a8d5a4b2cc65bf1517026b7db84fc6a94adb636d6f5b5c08d0e9778e83b" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515759 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-scripts\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515807 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515862 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4nzc\" (UniqueName: \"kubernetes.io/projected/1150a5ed-5385-4633-8806-664d91da5e86-kube-api-access-n4nzc\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515884 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-log-httpd\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515905 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515931 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-run-httpd\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515945 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-config-data\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.515969 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617484 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4nzc\" (UniqueName: \"kubernetes.io/projected/1150a5ed-5385-4633-8806-664d91da5e86-kube-api-access-n4nzc\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617532 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-log-httpd\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617558 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617587 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-run-httpd\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617605 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-config-data\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617629 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617651 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-scripts\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.617675 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.618173 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-run-httpd\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.618235 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-log-httpd\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.622115 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.622856 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.623082 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.624183 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-config-data\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.633190 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.636290 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-scripts\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.637032 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4nzc\" (UniqueName: \"kubernetes.io/projected/1150a5ed-5385-4633-8806-664d91da5e86-kube-api-access-n4nzc\") pod \"ceilometer-0\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.641881 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.774467 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.817611 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.922236 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f5vq\" (UniqueName: \"kubernetes.io/projected/33f82e98-788c-4584-bc61-3ed2704f00df-kube-api-access-6f5vq\") pod \"33f82e98-788c-4584-bc61-3ed2704f00df\" (UID: \"33f82e98-788c-4584-bc61-3ed2704f00df\") " Mar 14 09:52:04 crc kubenswrapper[4843]: I0314 09:52:04.947596 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33f82e98-788c-4584-bc61-3ed2704f00df-kube-api-access-6f5vq" (OuterVolumeSpecName: "kube-api-access-6f5vq") pod "33f82e98-788c-4584-bc61-3ed2704f00df" (UID: "33f82e98-788c-4584-bc61-3ed2704f00df"). InnerVolumeSpecName "kube-api-access-6f5vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.024166 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f5vq\" (UniqueName: \"kubernetes.io/projected/33f82e98-788c-4584-bc61-3ed2704f00df-kube-api-access-6f5vq\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.107817 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-fzhnw"] Mar 14 09:52:05 crc kubenswrapper[4843]: W0314 09:52:05.153305 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf6b7c42_02ea_4df8_a95f_9a2b3ae2e16b.slice/crio-1a3d627c2f042da5bd5db5406eaa73f48daeaa2fdd72199b6149908598529452 WatchSource:0}: Error finding container 1a3d627c2f042da5bd5db5406eaa73f48daeaa2fdd72199b6149908598529452: Status 404 returned error can't find the container with id 1a3d627c2f042da5bd5db5406eaa73f48daeaa2fdd72199b6149908598529452 Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.231021 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-25dc-account-create-update-qssln"] Mar 14 09:52:05 crc kubenswrapper[4843]: W0314 09:52:05.234326 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbef4afd5_c0a9_4443_aa51_358adb5b9014.slice/crio-e01b3af3b8857656426131a1c8ab8e3aa5b251abf9afea6eeaf34c74dd0cf4be WatchSource:0}: Error finding container e01b3af3b8857656426131a1c8ab8e3aa5b251abf9afea6eeaf34c74dd0cf4be: Status 404 returned error can't find the container with id e01b3af3b8857656426131a1c8ab8e3aa5b251abf9afea6eeaf34c74dd0cf4be Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.324695 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:05 crc kubenswrapper[4843]: W0314 09:52:05.326611 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1150a5ed_5385_4633_8806_664d91da5e86.slice/crio-9d16137c21fbd747dfa09158165add4dcd8ef693daadf2ee5368f9d16194636f WatchSource:0}: Error finding container 9d16137c21fbd747dfa09158165add4dcd8ef693daadf2ee5368f9d16194636f: Status 404 returned error can't find the container with id 9d16137c21fbd747dfa09158165add4dcd8ef693daadf2ee5368f9d16194636f Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.350007 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aebd0ba2-ff58-46a4-9ecf-1ada02952691" path="/var/lib/kubelet/pods/aebd0ba2-ff58-46a4-9ecf-1ada02952691/volumes" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.350691 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b55d408e-d1c2-4fe1-82b8-05e72c669dba" path="/var/lib/kubelet/pods/b55d408e-d1c2-4fe1-82b8-05e72c669dba/volumes" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.351219 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bba9218c-c199-4a70-bd95-f1c1de972639" path="/var/lib/kubelet/pods/bba9218c-c199-4a70-bd95-f1c1de972639/volumes" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.352344 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" event={"ID":"bef4afd5-c0a9-4443-aa51-358adb5b9014","Type":"ContainerStarted","Data":"e01b3af3b8857656426131a1c8ab8e3aa5b251abf9afea6eeaf34c74dd0cf4be"} Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.354554 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" event={"ID":"33f82e98-788c-4584-bc61-3ed2704f00df","Type":"ContainerDied","Data":"76028c178a7debbbfaa0d9e785a3b18444e5152052661fd6b81f246ab32a7473"} Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.354596 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76028c178a7debbbfaa0d9e785a3b18444e5152052661fd6b81f246ab32a7473" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.354881 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558032-ms2dx" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.361302 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerStarted","Data":"9d16137c21fbd747dfa09158165add4dcd8ef693daadf2ee5368f9d16194636f"} Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.363089 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-fzhnw" event={"ID":"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b","Type":"ContainerStarted","Data":"9b7aec10976a4de50e4011bf6809bc17b0b67d0a93693056f901317ec71a232f"} Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.363130 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-fzhnw" event={"ID":"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b","Type":"ContainerStarted","Data":"1a3d627c2f042da5bd5db5406eaa73f48daeaa2fdd72199b6149908598529452"} Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.380332 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-db-create-fzhnw" podStartSLOduration=1.380315113 podStartE2EDuration="1.380315113s" podCreationTimestamp="2026-03-14 09:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:05.378314525 +0000 UTC m=+2492.690925653" watchObservedRunningTime="2026-03-14 09:52:05.380315113 +0000 UTC m=+2492.692926241" Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.842934 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558026-fpp82"] Mar 14 09:52:05 crc kubenswrapper[4843]: I0314 09:52:05.858667 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558026-fpp82"] Mar 14 09:52:06 crc kubenswrapper[4843]: I0314 09:52:06.382789 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerStarted","Data":"0d6e352ba27ce611eaa7529af4be3db875ec1f8196dcb13e195517478fb4602f"} Mar 14 09:52:06 crc kubenswrapper[4843]: I0314 09:52:06.384204 4843 generic.go:334] "Generic (PLEG): container finished" podID="af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" containerID="9b7aec10976a4de50e4011bf6809bc17b0b67d0a93693056f901317ec71a232f" exitCode=0 Mar 14 09:52:06 crc kubenswrapper[4843]: I0314 09:52:06.384284 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-fzhnw" event={"ID":"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b","Type":"ContainerDied","Data":"9b7aec10976a4de50e4011bf6809bc17b0b67d0a93693056f901317ec71a232f"} Mar 14 09:52:06 crc kubenswrapper[4843]: I0314 09:52:06.385566 4843 generic.go:334] "Generic (PLEG): container finished" podID="bef4afd5-c0a9-4443-aa51-358adb5b9014" containerID="a462e3fb2aaec7845fb0dd1d1c07961c0e25ddaffc2ef3fde58df6337435f10d" exitCode=0 Mar 14 09:52:06 crc kubenswrapper[4843]: I0314 09:52:06.385598 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" event={"ID":"bef4afd5-c0a9-4443-aa51-358adb5b9014","Type":"ContainerDied","Data":"a462e3fb2aaec7845fb0dd1d1c07961c0e25ddaffc2ef3fde58df6337435f10d"} Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.359169 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44bd6132-462c-4296-9c39-c16070153e7c" path="/var/lib/kubelet/pods/44bd6132-462c-4296-9c39-c16070153e7c/volumes" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.401997 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerStarted","Data":"ca467eabcce50b68ec3b523ccc3f655f6d0247a3c96ae7d305126e698fc0f119"} Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.402066 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerStarted","Data":"99c8e9924270df7e804d2f684b6bfbd8acaf0dc887798821549f049127a721a6"} Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.878829 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.887442 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.977479 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef4afd5-c0a9-4443-aa51-358adb5b9014-operator-scripts\") pod \"bef4afd5-c0a9-4443-aa51-358adb5b9014\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.977644 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-operator-scripts\") pod \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.978352 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef4afd5-c0a9-4443-aa51-358adb5b9014-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bef4afd5-c0a9-4443-aa51-358adb5b9014" (UID: "bef4afd5-c0a9-4443-aa51-358adb5b9014"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.978433 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" (UID: "af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.978526 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqsmx\" (UniqueName: \"kubernetes.io/projected/bef4afd5-c0a9-4443-aa51-358adb5b9014-kube-api-access-pqsmx\") pod \"bef4afd5-c0a9-4443-aa51-358adb5b9014\" (UID: \"bef4afd5-c0a9-4443-aa51-358adb5b9014\") " Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.978965 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cclg\" (UniqueName: \"kubernetes.io/projected/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-kube-api-access-6cclg\") pod \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\" (UID: \"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b\") " Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.979763 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef4afd5-c0a9-4443-aa51-358adb5b9014-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.979781 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.990191 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-kube-api-access-6cclg" (OuterVolumeSpecName: "kube-api-access-6cclg") pod "af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" (UID: "af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b"). InnerVolumeSpecName "kube-api-access-6cclg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:07 crc kubenswrapper[4843]: I0314 09:52:07.990235 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef4afd5-c0a9-4443-aa51-358adb5b9014-kube-api-access-pqsmx" (OuterVolumeSpecName: "kube-api-access-pqsmx") pod "bef4afd5-c0a9-4443-aa51-358adb5b9014" (UID: "bef4afd5-c0a9-4443-aa51-358adb5b9014"). InnerVolumeSpecName "kube-api-access-pqsmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.081178 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqsmx\" (UniqueName: \"kubernetes.io/projected/bef4afd5-c0a9-4443-aa51-358adb5b9014-kube-api-access-pqsmx\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.081212 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cclg\" (UniqueName: \"kubernetes.io/projected/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b-kube-api-access-6cclg\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.414915 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-fzhnw" event={"ID":"af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b","Type":"ContainerDied","Data":"1a3d627c2f042da5bd5db5406eaa73f48daeaa2fdd72199b6149908598529452"} Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.414958 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a3d627c2f042da5bd5db5406eaa73f48daeaa2fdd72199b6149908598529452" Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.415022 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-fzhnw" Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.419009 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" event={"ID":"bef4afd5-c0a9-4443-aa51-358adb5b9014","Type":"ContainerDied","Data":"e01b3af3b8857656426131a1c8ab8e3aa5b251abf9afea6eeaf34c74dd0cf4be"} Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.419051 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01b3af3b8857656426131a1c8ab8e3aa5b251abf9afea6eeaf34c74dd0cf4be" Mar 14 09:52:08 crc kubenswrapper[4843]: I0314 09:52:08.419112 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-25dc-account-create-update-qssln" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.449369 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerStarted","Data":"abddf69c9b878baf399a0a5521e31f5325289564a81c006a1f18401efea403ce"} Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.449972 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.489855 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.013254002 podStartE2EDuration="5.489827815s" podCreationTimestamp="2026-03-14 09:52:04 +0000 UTC" firstStartedPulling="2026-03-14 09:52:05.331200101 +0000 UTC m=+2492.643811219" lastFinishedPulling="2026-03-14 09:52:08.807773904 +0000 UTC m=+2496.120385032" observedRunningTime="2026-03-14 09:52:09.478756354 +0000 UTC m=+2496.791367482" watchObservedRunningTime="2026-03-14 09:52:09.489827815 +0000 UTC m=+2496.802438973" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.718039 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw"] Mar 14 09:52:09 crc kubenswrapper[4843]: E0314 09:52:09.718789 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f82e98-788c-4584-bc61-3ed2704f00df" containerName="oc" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.718877 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f82e98-788c-4584-bc61-3ed2704f00df" containerName="oc" Mar 14 09:52:09 crc kubenswrapper[4843]: E0314 09:52:09.718969 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef4afd5-c0a9-4443-aa51-358adb5b9014" containerName="mariadb-account-create-update" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.719042 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef4afd5-c0a9-4443-aa51-358adb5b9014" containerName="mariadb-account-create-update" Mar 14 09:52:09 crc kubenswrapper[4843]: E0314 09:52:09.719133 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" containerName="mariadb-database-create" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.719194 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" containerName="mariadb-database-create" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.719452 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f82e98-788c-4584-bc61-3ed2704f00df" containerName="oc" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.719549 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" containerName="mariadb-database-create" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.719639 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef4afd5-c0a9-4443-aa51-358adb5b9014" containerName="mariadb-account-create-update" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.720463 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.723357 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.723554 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-9dllr" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.731619 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw"] Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.809891 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-config-data\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.809943 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7bzg\" (UniqueName: \"kubernetes.io/projected/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-kube-api-access-x7bzg\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.810026 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-db-sync-config-data\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.810122 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.911733 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-config-data\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.911801 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7bzg\" (UniqueName: \"kubernetes.io/projected/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-kube-api-access-x7bzg\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.911849 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-db-sync-config-data\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.911923 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.917085 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-config-data\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.918636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-db-sync-config-data\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.926400 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:09 crc kubenswrapper[4843]: I0314 09:52:09.928465 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7bzg\" (UniqueName: \"kubernetes.io/projected/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-kube-api-access-x7bzg\") pod \"watcher-kuttl-db-sync-tqfrw\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:10 crc kubenswrapper[4843]: I0314 09:52:10.037497 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:10 crc kubenswrapper[4843]: W0314 09:52:10.576046 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77b71daa_ca7f_4f50_8c03_31d4fd4ba2e0.slice/crio-9e035e0075497449c6f4ecbb1cdfb554f6638fe57286595617ff5dc7253de51d WatchSource:0}: Error finding container 9e035e0075497449c6f4ecbb1cdfb554f6638fe57286595617ff5dc7253de51d: Status 404 returned error can't find the container with id 9e035e0075497449c6f4ecbb1cdfb554f6638fe57286595617ff5dc7253de51d Mar 14 09:52:10 crc kubenswrapper[4843]: I0314 09:52:10.578966 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw"] Mar 14 09:52:11 crc kubenswrapper[4843]: I0314 09:52:11.466908 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" event={"ID":"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0","Type":"ContainerStarted","Data":"3d6c1c7d106d7cfdf995dae9c299e63b33e1be9f07850d6d6ae1a85d4c9b2a3d"} Mar 14 09:52:11 crc kubenswrapper[4843]: I0314 09:52:11.467236 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" event={"ID":"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0","Type":"ContainerStarted","Data":"9e035e0075497449c6f4ecbb1cdfb554f6638fe57286595617ff5dc7253de51d"} Mar 14 09:52:11 crc kubenswrapper[4843]: I0314 09:52:11.494913 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" podStartSLOduration=2.494895044 podStartE2EDuration="2.494895044s" podCreationTimestamp="2026-03-14 09:52:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:11.494593327 +0000 UTC m=+2498.807204455" watchObservedRunningTime="2026-03-14 09:52:11.494895044 +0000 UTC m=+2498.807506172" Mar 14 09:52:13 crc kubenswrapper[4843]: I0314 09:52:13.487751 4843 generic.go:334] "Generic (PLEG): container finished" podID="77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" containerID="3d6c1c7d106d7cfdf995dae9c299e63b33e1be9f07850d6d6ae1a85d4c9b2a3d" exitCode=0 Mar 14 09:52:13 crc kubenswrapper[4843]: I0314 09:52:13.487858 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" event={"ID":"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0","Type":"ContainerDied","Data":"3d6c1c7d106d7cfdf995dae9c299e63b33e1be9f07850d6d6ae1a85d4c9b2a3d"} Mar 14 09:52:14 crc kubenswrapper[4843]: I0314 09:52:14.338708 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:52:14 crc kubenswrapper[4843]: E0314 09:52:14.339120 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:52:14 crc kubenswrapper[4843]: I0314 09:52:14.940846 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:14 crc kubenswrapper[4843]: I0314 09:52:14.996390 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-combined-ca-bundle\") pod \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " Mar 14 09:52:14 crc kubenswrapper[4843]: I0314 09:52:14.996468 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-config-data\") pod \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " Mar 14 09:52:14 crc kubenswrapper[4843]: I0314 09:52:14.996500 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-db-sync-config-data\") pod \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " Mar 14 09:52:14 crc kubenswrapper[4843]: I0314 09:52:14.996547 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7bzg\" (UniqueName: \"kubernetes.io/projected/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-kube-api-access-x7bzg\") pod \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\" (UID: \"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0\") " Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.005081 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-kube-api-access-x7bzg" (OuterVolumeSpecName: "kube-api-access-x7bzg") pod "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" (UID: "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0"). InnerVolumeSpecName "kube-api-access-x7bzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.036432 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" (UID: "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.040233 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" (UID: "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.044410 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-config-data" (OuterVolumeSpecName: "config-data") pod "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" (UID: "77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.098998 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.099043 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.099059 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7bzg\" (UniqueName: \"kubernetes.io/projected/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-kube-api-access-x7bzg\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.099071 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.506795 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" event={"ID":"77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0","Type":"ContainerDied","Data":"9e035e0075497449c6f4ecbb1cdfb554f6638fe57286595617ff5dc7253de51d"} Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.506837 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e035e0075497449c6f4ecbb1cdfb554f6638fe57286595617ff5dc7253de51d" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.506895 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.810117 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:15 crc kubenswrapper[4843]: E0314 09:52:15.810483 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" containerName="watcher-kuttl-db-sync" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.810500 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" containerName="watcher-kuttl-db-sync" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.810640 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" containerName="watcher-kuttl-db-sync" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.811167 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.813322 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.813603 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-9dllr" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.821393 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.835897 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.837155 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.847931 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.848805 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.867883 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.868948 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.874227 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.904131 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913027 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m28nn\" (UniqueName: \"kubernetes.io/projected/b2025e74-2890-43c4-a656-e0f9eb2db5dd-kube-api-access-m28nn\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913092 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913127 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913151 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913175 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2025e74-2890-43c4-a656-e0f9eb2db5dd-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913197 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d11aac1-e789-4561-b182-1e6f150f10ac-logs\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913240 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913261 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913306 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913334 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p5r4\" (UniqueName: \"kubernetes.io/projected/9d11aac1-e789-4561-b182-1e6f150f10ac-kube-api-access-2p5r4\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:15 crc kubenswrapper[4843]: I0314 09:52:15.913381 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015264 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015320 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015350 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015376 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015397 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2025e74-2890-43c4-a656-e0f9eb2db5dd-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015415 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d11aac1-e789-4561-b182-1e6f150f10ac-logs\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015431 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015450 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015464 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015486 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015506 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015527 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7jlj\" (UniqueName: \"kubernetes.io/projected/282f4b34-3137-454f-8733-9d2e8cccb88f-kube-api-access-z7jlj\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015544 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p5r4\" (UniqueName: \"kubernetes.io/projected/9d11aac1-e789-4561-b182-1e6f150f10ac-kube-api-access-2p5r4\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015576 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015619 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/282f4b34-3137-454f-8733-9d2e8cccb88f-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015646 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.015667 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m28nn\" (UniqueName: \"kubernetes.io/projected/b2025e74-2890-43c4-a656-e0f9eb2db5dd-kube-api-access-m28nn\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.016966 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2025e74-2890-43c4-a656-e0f9eb2db5dd-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.017238 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d11aac1-e789-4561-b182-1e6f150f10ac-logs\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.020360 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.020911 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.021332 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.021828 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.022350 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.023995 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.035571 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.036938 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m28nn\" (UniqueName: \"kubernetes.io/projected/b2025e74-2890-43c4-a656-e0f9eb2db5dd-kube-api-access-m28nn\") pod \"watcher-kuttl-applier-0\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.047730 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p5r4\" (UniqueName: \"kubernetes.io/projected/9d11aac1-e789-4561-b182-1e6f150f10ac-kube-api-access-2p5r4\") pod \"watcher-kuttl-api-0\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.116865 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.116928 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.116958 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.116984 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7jlj\" (UniqueName: \"kubernetes.io/projected/282f4b34-3137-454f-8733-9d2e8cccb88f-kube-api-access-z7jlj\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.117047 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/282f4b34-3137-454f-8733-9d2e8cccb88f-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.117076 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.117658 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/282f4b34-3137-454f-8733-9d2e8cccb88f-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.120473 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.120617 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.120874 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.120989 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.128753 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.147147 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7jlj\" (UniqueName: \"kubernetes.io/projected/282f4b34-3137-454f-8733-9d2e8cccb88f-kube-api-access-z7jlj\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.174752 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.195345 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.774035 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.851618 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:16 crc kubenswrapper[4843]: I0314 09:52:16.863512 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:16 crc kubenswrapper[4843]: W0314 09:52:16.874478 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod282f4b34_3137_454f_8733_9d2e8cccb88f.slice/crio-96ffae1133e1a016b06748d068f9fb14e367f9ad7556dac744dc8983fd1585b1 WatchSource:0}: Error finding container 96ffae1133e1a016b06748d068f9fb14e367f9ad7556dac744dc8983fd1585b1: Status 404 returned error can't find the container with id 96ffae1133e1a016b06748d068f9fb14e367f9ad7556dac744dc8983fd1585b1 Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.527697 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"9d11aac1-e789-4561-b182-1e6f150f10ac","Type":"ContainerStarted","Data":"eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.527973 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"9d11aac1-e789-4561-b182-1e6f150f10ac","Type":"ContainerStarted","Data":"a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.527984 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"9d11aac1-e789-4561-b182-1e6f150f10ac","Type":"ContainerStarted","Data":"5d121f9f7beb30b6e3b518e445fcf604158144f76f7a67021ab4ea5ea3ce5de8"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.529026 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.532758 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2025e74-2890-43c4-a656-e0f9eb2db5dd","Type":"ContainerStarted","Data":"d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.532807 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2025e74-2890-43c4-a656-e0f9eb2db5dd","Type":"ContainerStarted","Data":"59f4e6d0913b27b26b60bedea31b9ba31a18874ad4f755b67010574ffc17594c"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.536021 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"282f4b34-3137-454f-8733-9d2e8cccb88f","Type":"ContainerStarted","Data":"047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.536062 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"282f4b34-3137-454f-8733-9d2e8cccb88f","Type":"ContainerStarted","Data":"96ffae1133e1a016b06748d068f9fb14e367f9ad7556dac744dc8983fd1585b1"} Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.550358 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.5503405089999998 podStartE2EDuration="2.550340509s" podCreationTimestamp="2026-03-14 09:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:17.547933669 +0000 UTC m=+2504.860544797" watchObservedRunningTime="2026-03-14 09:52:17.550340509 +0000 UTC m=+2504.862951637" Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.565696 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.565675584 podStartE2EDuration="2.565675584s" podCreationTimestamp="2026-03-14 09:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:17.564453264 +0000 UTC m=+2504.877064392" watchObservedRunningTime="2026-03-14 09:52:17.565675584 +0000 UTC m=+2504.878286712" Mar 14 09:52:17 crc kubenswrapper[4843]: I0314 09:52:17.586305 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.586286819 podStartE2EDuration="2.586286819s" podCreationTimestamp="2026-03-14 09:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:17.57977709 +0000 UTC m=+2504.892388218" watchObservedRunningTime="2026-03-14 09:52:17.586286819 +0000 UTC m=+2504.898897947" Mar 14 09:52:19 crc kubenswrapper[4843]: I0314 09:52:19.555258 4843 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 14 09:52:19 crc kubenswrapper[4843]: I0314 09:52:19.953078 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:21 crc kubenswrapper[4843]: I0314 09:52:21.129859 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:21 crc kubenswrapper[4843]: I0314 09:52:21.211660 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.129124 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.156569 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.176264 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.182510 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.195838 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.222070 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.618251 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.624993 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.658121 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:26 crc kubenswrapper[4843]: I0314 09:52:26.669114 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.102616 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.103314 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-central-agent" containerID="cri-o://0d6e352ba27ce611eaa7529af4be3db875ec1f8196dcb13e195517478fb4602f" gracePeriod=30 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.103352 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="sg-core" containerID="cri-o://ca467eabcce50b68ec3b523ccc3f655f6d0247a3c96ae7d305126e698fc0f119" gracePeriod=30 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.103417 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="proxy-httpd" containerID="cri-o://abddf69c9b878baf399a0a5521e31f5325289564a81c006a1f18401efea403ce" gracePeriod=30 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.103457 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-notification-agent" containerID="cri-o://99c8e9924270df7e804d2f684b6bfbd8acaf0dc887798821549f049127a721a6" gracePeriod=30 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.109789 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.219:3000/\": read tcp 10.217.0.2:39730->10.217.0.219:3000: read: connection reset by peer" Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.338672 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:52:29 crc kubenswrapper[4843]: E0314 09:52:29.339195 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.641896 4843 generic.go:334] "Generic (PLEG): container finished" podID="1150a5ed-5385-4633-8806-664d91da5e86" containerID="abddf69c9b878baf399a0a5521e31f5325289564a81c006a1f18401efea403ce" exitCode=0 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.641930 4843 generic.go:334] "Generic (PLEG): container finished" podID="1150a5ed-5385-4633-8806-664d91da5e86" containerID="ca467eabcce50b68ec3b523ccc3f655f6d0247a3c96ae7d305126e698fc0f119" exitCode=2 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.641940 4843 generic.go:334] "Generic (PLEG): container finished" podID="1150a5ed-5385-4633-8806-664d91da5e86" containerID="99c8e9924270df7e804d2f684b6bfbd8acaf0dc887798821549f049127a721a6" exitCode=0 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.641949 4843 generic.go:334] "Generic (PLEG): container finished" podID="1150a5ed-5385-4633-8806-664d91da5e86" containerID="0d6e352ba27ce611eaa7529af4be3db875ec1f8196dcb13e195517478fb4602f" exitCode=0 Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.641964 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerDied","Data":"abddf69c9b878baf399a0a5521e31f5325289564a81c006a1f18401efea403ce"} Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.641997 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerDied","Data":"ca467eabcce50b68ec3b523ccc3f655f6d0247a3c96ae7d305126e698fc0f119"} Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.642008 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerDied","Data":"99c8e9924270df7e804d2f684b6bfbd8acaf0dc887798821549f049127a721a6"} Mar 14 09:52:29 crc kubenswrapper[4843]: I0314 09:52:29.642018 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerDied","Data":"0d6e352ba27ce611eaa7529af4be3db875ec1f8196dcb13e195517478fb4602f"} Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.010478 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.067731 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-scripts\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.068523 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4nzc\" (UniqueName: \"kubernetes.io/projected/1150a5ed-5385-4633-8806-664d91da5e86-kube-api-access-n4nzc\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.068720 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-log-httpd\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.068854 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-ceilometer-tls-certs\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.068938 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-run-httpd\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.069013 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-sg-core-conf-yaml\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.069094 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-combined-ca-bundle\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.069175 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-config-data\") pod \"1150a5ed-5385-4633-8806-664d91da5e86\" (UID: \"1150a5ed-5385-4633-8806-664d91da5e86\") " Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.069511 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.069537 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.074027 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1150a5ed-5385-4633-8806-664d91da5e86-kube-api-access-n4nzc" (OuterVolumeSpecName: "kube-api-access-n4nzc") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "kube-api-access-n4nzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.074165 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-scripts" (OuterVolumeSpecName: "scripts") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.092982 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.112954 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.129663 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.153475 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-config-data" (OuterVolumeSpecName: "config-data") pod "1150a5ed-5385-4633-8806-664d91da5e86" (UID: "1150a5ed-5385-4633-8806-664d91da5e86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171564 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171601 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4nzc\" (UniqueName: \"kubernetes.io/projected/1150a5ed-5385-4633-8806-664d91da5e86-kube-api-access-n4nzc\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171613 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171625 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171637 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1150a5ed-5385-4633-8806-664d91da5e86-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171649 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171660 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.171670 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1150a5ed-5385-4633-8806-664d91da5e86-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.652565 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"1150a5ed-5385-4633-8806-664d91da5e86","Type":"ContainerDied","Data":"9d16137c21fbd747dfa09158165add4dcd8ef693daadf2ee5368f9d16194636f"} Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.652627 4843 scope.go:117] "RemoveContainer" containerID="abddf69c9b878baf399a0a5521e31f5325289564a81c006a1f18401efea403ce" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.652798 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.671180 4843 scope.go:117] "RemoveContainer" containerID="ca467eabcce50b68ec3b523ccc3f655f6d0247a3c96ae7d305126e698fc0f119" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.691139 4843 scope.go:117] "RemoveContainer" containerID="99c8e9924270df7e804d2f684b6bfbd8acaf0dc887798821549f049127a721a6" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.703356 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.710238 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.730505 4843 scope.go:117] "RemoveContainer" containerID="0d6e352ba27ce611eaa7529af4be3db875ec1f8196dcb13e195517478fb4602f" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.736538 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:30 crc kubenswrapper[4843]: E0314 09:52:30.736967 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="sg-core" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.736981 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="sg-core" Mar 14 09:52:30 crc kubenswrapper[4843]: E0314 09:52:30.736998 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="proxy-httpd" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737005 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="proxy-httpd" Mar 14 09:52:30 crc kubenswrapper[4843]: E0314 09:52:30.737034 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-central-agent" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737042 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-central-agent" Mar 14 09:52:30 crc kubenswrapper[4843]: E0314 09:52:30.737054 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-notification-agent" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737062 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-notification-agent" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737273 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="proxy-httpd" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737292 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-central-agent" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737353 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="ceilometer-notification-agent" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.737369 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="1150a5ed-5385-4633-8806-664d91da5e86" containerName="sg-core" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.739094 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.741062 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.741411 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.741534 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.743942 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.888432 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.888885 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.888922 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-log-httpd\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.889036 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.889122 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-scripts\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.889145 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-config-data\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.889311 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-run-httpd\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.889364 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szz9l\" (UniqueName: \"kubernetes.io/projected/53c8bb85-a064-4ca4-a541-79be614e88bd-kube-api-access-szz9l\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990453 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-scripts\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990495 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-config-data\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990535 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-run-httpd\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990561 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szz9l\" (UniqueName: \"kubernetes.io/projected/53c8bb85-a064-4ca4-a541-79be614e88bd-kube-api-access-szz9l\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990593 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990620 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990639 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-log-httpd\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.990674 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.991526 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-run-httpd\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.991851 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-log-httpd\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:30 crc kubenswrapper[4843]: I0314 09:52:30.995850 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.007352 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-config-data\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.007670 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-scripts\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.008017 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.009537 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szz9l\" (UniqueName: \"kubernetes.io/projected/53c8bb85-a064-4ca4-a541-79be614e88bd-kube-api-access-szz9l\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.010239 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.084960 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.356626 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1150a5ed-5385-4633-8806-664d91da5e86" path="/var/lib/kubelet/pods/1150a5ed-5385-4633-8806-664d91da5e86/volumes" Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.587221 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:31 crc kubenswrapper[4843]: I0314 09:52:31.664156 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerStarted","Data":"fa12dbfd7723f88e4098104e819d248115c99a107bbfa0fb712d9d813a0878aa"} Mar 14 09:52:32 crc kubenswrapper[4843]: I0314 09:52:32.677336 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerStarted","Data":"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983"} Mar 14 09:52:33 crc kubenswrapper[4843]: I0314 09:52:33.690467 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerStarted","Data":"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70"} Mar 14 09:52:33 crc kubenswrapper[4843]: I0314 09:52:33.690810 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerStarted","Data":"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b"} Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.709518 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerStarted","Data":"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e"} Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.710093 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.737632 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.746427 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-tqfrw"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.762749 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.168235385 podStartE2EDuration="5.762728767s" podCreationTimestamp="2026-03-14 09:52:30 +0000 UTC" firstStartedPulling="2026-03-14 09:52:31.578806732 +0000 UTC m=+2518.891417860" lastFinishedPulling="2026-03-14 09:52:35.173300114 +0000 UTC m=+2522.485911242" observedRunningTime="2026-03-14 09:52:35.756661818 +0000 UTC m=+2523.069272956" watchObservedRunningTime="2026-03-14 09:52:35.762728767 +0000 UTC m=+2523.075339885" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.788602 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher25dc-account-delete-cr72d"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.789720 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.799619 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher25dc-account-delete-cr72d"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.831103 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.831376 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" containerName="watcher-applier" containerID="cri-o://d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19" gracePeriod=30 Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.872413 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9qrg\" (UniqueName: \"kubernetes.io/projected/286e0246-8d7d-47e9-a44b-3d7acb0ec046-kube-api-access-f9qrg\") pod \"watcher25dc-account-delete-cr72d\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.872794 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/286e0246-8d7d-47e9-a44b-3d7acb0ec046-operator-scripts\") pod \"watcher25dc-account-delete-cr72d\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.893180 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.893553 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="282f4b34-3137-454f-8733-9d2e8cccb88f" containerName="watcher-decision-engine" containerID="cri-o://047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff" gracePeriod=30 Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.944895 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.945135 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-kuttl-api-log" containerID="cri-o://a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e" gracePeriod=30 Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.945281 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-api" containerID="cri-o://eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba" gracePeriod=30 Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.974008 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9qrg\" (UniqueName: \"kubernetes.io/projected/286e0246-8d7d-47e9-a44b-3d7acb0ec046-kube-api-access-f9qrg\") pod \"watcher25dc-account-delete-cr72d\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.974153 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/286e0246-8d7d-47e9-a44b-3d7acb0ec046-operator-scripts\") pod \"watcher25dc-account-delete-cr72d\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:35 crc kubenswrapper[4843]: I0314 09:52:35.975098 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/286e0246-8d7d-47e9-a44b-3d7acb0ec046-operator-scripts\") pod \"watcher25dc-account-delete-cr72d\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.001876 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9qrg\" (UniqueName: \"kubernetes.io/projected/286e0246-8d7d-47e9-a44b-3d7acb0ec046-kube-api-access-f9qrg\") pod \"watcher25dc-account-delete-cr72d\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.115651 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:36 crc kubenswrapper[4843]: E0314 09:52:36.131644 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:52:36 crc kubenswrapper[4843]: E0314 09:52:36.133181 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:52:36 crc kubenswrapper[4843]: E0314 09:52:36.134421 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:52:36 crc kubenswrapper[4843]: E0314 09:52:36.134459 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" containerName="watcher-applier" Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.667983 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher25dc-account-delete-cr72d"] Mar 14 09:52:36 crc kubenswrapper[4843]: W0314 09:52:36.672448 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod286e0246_8d7d_47e9_a44b_3d7acb0ec046.slice/crio-007a0e940fd5b2c7311f1ccae7729f741387bc6430978ebae45660051d1b8115 WatchSource:0}: Error finding container 007a0e940fd5b2c7311f1ccae7729f741387bc6430978ebae45660051d1b8115: Status 404 returned error can't find the container with id 007a0e940fd5b2c7311f1ccae7729f741387bc6430978ebae45660051d1b8115 Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.725140 4843 generic.go:334] "Generic (PLEG): container finished" podID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerID="a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e" exitCode=143 Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.725248 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"9d11aac1-e789-4561-b182-1e6f150f10ac","Type":"ContainerDied","Data":"a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e"} Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.727251 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" event={"ID":"286e0246-8d7d-47e9-a44b-3d7acb0ec046","Type":"ContainerStarted","Data":"007a0e940fd5b2c7311f1ccae7729f741387bc6430978ebae45660051d1b8115"} Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.812102 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.222:9322/\": read tcp 10.217.0.2:41180->10.217.0.222:9322: read: connection reset by peer" Mar 14 09:52:36 crc kubenswrapper[4843]: I0314 09:52:36.812207 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.0.222:9322/\": read tcp 10.217.0.2:41178->10.217.0.222:9322: read: connection reset by peer" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.216753 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.303848 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-config-data\") pod \"9d11aac1-e789-4561-b182-1e6f150f10ac\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.303917 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p5r4\" (UniqueName: \"kubernetes.io/projected/9d11aac1-e789-4561-b182-1e6f150f10ac-kube-api-access-2p5r4\") pod \"9d11aac1-e789-4561-b182-1e6f150f10ac\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.303966 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d11aac1-e789-4561-b182-1e6f150f10ac-logs\") pod \"9d11aac1-e789-4561-b182-1e6f150f10ac\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.304005 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-combined-ca-bundle\") pod \"9d11aac1-e789-4561-b182-1e6f150f10ac\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.304072 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-cert-memcached-mtls\") pod \"9d11aac1-e789-4561-b182-1e6f150f10ac\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.304164 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-custom-prometheus-ca\") pod \"9d11aac1-e789-4561-b182-1e6f150f10ac\" (UID: \"9d11aac1-e789-4561-b182-1e6f150f10ac\") " Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.305587 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d11aac1-e789-4561-b182-1e6f150f10ac-logs" (OuterVolumeSpecName: "logs") pod "9d11aac1-e789-4561-b182-1e6f150f10ac" (UID: "9d11aac1-e789-4561-b182-1e6f150f10ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.305705 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d11aac1-e789-4561-b182-1e6f150f10ac-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.330043 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d11aac1-e789-4561-b182-1e6f150f10ac-kube-api-access-2p5r4" (OuterVolumeSpecName: "kube-api-access-2p5r4") pod "9d11aac1-e789-4561-b182-1e6f150f10ac" (UID: "9d11aac1-e789-4561-b182-1e6f150f10ac"). InnerVolumeSpecName "kube-api-access-2p5r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.337023 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d11aac1-e789-4561-b182-1e6f150f10ac" (UID: "9d11aac1-e789-4561-b182-1e6f150f10ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.372159 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-config-data" (OuterVolumeSpecName: "config-data") pod "9d11aac1-e789-4561-b182-1e6f150f10ac" (UID: "9d11aac1-e789-4561-b182-1e6f150f10ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.375433 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0" path="/var/lib/kubelet/pods/77b71daa-ca7f-4f50-8c03-31d4fd4ba2e0/volumes" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.381410 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "9d11aac1-e789-4561-b182-1e6f150f10ac" (UID: "9d11aac1-e789-4561-b182-1e6f150f10ac"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.408458 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.408500 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p5r4\" (UniqueName: \"kubernetes.io/projected/9d11aac1-e789-4561-b182-1e6f150f10ac-kube-api-access-2p5r4\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.408516 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.408529 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.457620 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "9d11aac1-e789-4561-b182-1e6f150f10ac" (UID: "9d11aac1-e789-4561-b182-1e6f150f10ac"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.510034 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/9d11aac1-e789-4561-b182-1e6f150f10ac-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.736937 4843 generic.go:334] "Generic (PLEG): container finished" podID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerID="eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba" exitCode=0 Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.736994 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"9d11aac1-e789-4561-b182-1e6f150f10ac","Type":"ContainerDied","Data":"eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba"} Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.737001 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.737019 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"9d11aac1-e789-4561-b182-1e6f150f10ac","Type":"ContainerDied","Data":"5d121f9f7beb30b6e3b518e445fcf604158144f76f7a67021ab4ea5ea3ce5de8"} Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.737038 4843 scope.go:117] "RemoveContainer" containerID="eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.739467 4843 generic.go:334] "Generic (PLEG): container finished" podID="286e0246-8d7d-47e9-a44b-3d7acb0ec046" containerID="9c5ec8b64bee09261423f38bef854d84055bbef2c07bcb65e7c3297828dd988b" exitCode=0 Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.739506 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" event={"ID":"286e0246-8d7d-47e9-a44b-3d7acb0ec046","Type":"ContainerDied","Data":"9c5ec8b64bee09261423f38bef854d84055bbef2c07bcb65e7c3297828dd988b"} Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.766883 4843 scope.go:117] "RemoveContainer" containerID="a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.784088 4843 scope.go:117] "RemoveContainer" containerID="eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba" Mar 14 09:52:37 crc kubenswrapper[4843]: E0314 09:52:37.784727 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba\": container with ID starting with eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba not found: ID does not exist" containerID="eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.784763 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba"} err="failed to get container status \"eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba\": rpc error: code = NotFound desc = could not find container \"eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba\": container with ID starting with eba3180bb68155a7ca0ab0b91e690437cbe7f3650a7f25b5ffc8d310f38157ba not found: ID does not exist" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.784807 4843 scope.go:117] "RemoveContainer" containerID="a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e" Mar 14 09:52:37 crc kubenswrapper[4843]: E0314 09:52:37.785223 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e\": container with ID starting with a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e not found: ID does not exist" containerID="a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.785246 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e"} err="failed to get container status \"a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e\": rpc error: code = NotFound desc = could not find container \"a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e\": container with ID starting with a47f3d733e15b8d80860d28019ce0551c9b75bd08849c3422d81f6cb2a896d5e not found: ID does not exist" Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.785285 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:37 crc kubenswrapper[4843]: I0314 09:52:37.792671 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.753612 4843 generic.go:334] "Generic (PLEG): container finished" podID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" containerID="d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19" exitCode=0 Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.753722 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2025e74-2890-43c4-a656-e0f9eb2db5dd","Type":"ContainerDied","Data":"d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19"} Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.905122 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.905442 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-central-agent" containerID="cri-o://84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" gracePeriod=30 Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.906170 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="proxy-httpd" containerID="cri-o://1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" gracePeriod=30 Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.906235 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="sg-core" containerID="cri-o://2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" gracePeriod=30 Mar 14 09:52:38 crc kubenswrapper[4843]: I0314 09:52:38.907984 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-notification-agent" containerID="cri-o://206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" gracePeriod=30 Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.088092 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.149375 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2025e74-2890-43c4-a656-e0f9eb2db5dd-logs\") pod \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.149454 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-combined-ca-bundle\") pod \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.149577 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-cert-memcached-mtls\") pod \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.149638 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m28nn\" (UniqueName: \"kubernetes.io/projected/b2025e74-2890-43c4-a656-e0f9eb2db5dd-kube-api-access-m28nn\") pod \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.149718 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-config-data\") pod \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\" (UID: \"b2025e74-2890-43c4-a656-e0f9eb2db5dd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.149873 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2025e74-2890-43c4-a656-e0f9eb2db5dd-logs" (OuterVolumeSpecName: "logs") pod "b2025e74-2890-43c4-a656-e0f9eb2db5dd" (UID: "b2025e74-2890-43c4-a656-e0f9eb2db5dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.150135 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2025e74-2890-43c4-a656-e0f9eb2db5dd-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.161964 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2025e74-2890-43c4-a656-e0f9eb2db5dd-kube-api-access-m28nn" (OuterVolumeSpecName: "kube-api-access-m28nn") pod "b2025e74-2890-43c4-a656-e0f9eb2db5dd" (UID: "b2025e74-2890-43c4-a656-e0f9eb2db5dd"). InnerVolumeSpecName "kube-api-access-m28nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.193167 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2025e74-2890-43c4-a656-e0f9eb2db5dd" (UID: "b2025e74-2890-43c4-a656-e0f9eb2db5dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.212059 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-config-data" (OuterVolumeSpecName: "config-data") pod "b2025e74-2890-43c4-a656-e0f9eb2db5dd" (UID: "b2025e74-2890-43c4-a656-e0f9eb2db5dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.226852 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "b2025e74-2890-43c4-a656-e0f9eb2db5dd" (UID: "b2025e74-2890-43c4-a656-e0f9eb2db5dd"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.251794 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.251864 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m28nn\" (UniqueName: \"kubernetes.io/projected/b2025e74-2890-43c4-a656-e0f9eb2db5dd-kube-api-access-m28nn\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.251881 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.251892 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2025e74-2890-43c4-a656-e0f9eb2db5dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.277887 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.352763 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/286e0246-8d7d-47e9-a44b-3d7acb0ec046-operator-scripts\") pod \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.352919 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9qrg\" (UniqueName: \"kubernetes.io/projected/286e0246-8d7d-47e9-a44b-3d7acb0ec046-kube-api-access-f9qrg\") pod \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\" (UID: \"286e0246-8d7d-47e9-a44b-3d7acb0ec046\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.354086 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286e0246-8d7d-47e9-a44b-3d7acb0ec046-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "286e0246-8d7d-47e9-a44b-3d7acb0ec046" (UID: "286e0246-8d7d-47e9-a44b-3d7acb0ec046"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.357147 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286e0246-8d7d-47e9-a44b-3d7acb0ec046-kube-api-access-f9qrg" (OuterVolumeSpecName: "kube-api-access-f9qrg") pod "286e0246-8d7d-47e9-a44b-3d7acb0ec046" (UID: "286e0246-8d7d-47e9-a44b-3d7acb0ec046"). InnerVolumeSpecName "kube-api-access-f9qrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.358249 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" path="/var/lib/kubelet/pods/9d11aac1-e789-4561-b182-1e6f150f10ac/volumes" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.455104 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9qrg\" (UniqueName: \"kubernetes.io/projected/286e0246-8d7d-47e9-a44b-3d7acb0ec046-kube-api-access-f9qrg\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.455145 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/286e0246-8d7d-47e9-a44b-3d7acb0ec046-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.689404 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.759889 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-cert-memcached-mtls\") pod \"282f4b34-3137-454f-8733-9d2e8cccb88f\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.759985 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-combined-ca-bundle\") pod \"282f4b34-3137-454f-8733-9d2e8cccb88f\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.760013 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-config-data\") pod \"282f4b34-3137-454f-8733-9d2e8cccb88f\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.760071 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7jlj\" (UniqueName: \"kubernetes.io/projected/282f4b34-3137-454f-8733-9d2e8cccb88f-kube-api-access-z7jlj\") pod \"282f4b34-3137-454f-8733-9d2e8cccb88f\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.760148 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-custom-prometheus-ca\") pod \"282f4b34-3137-454f-8733-9d2e8cccb88f\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.760185 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/282f4b34-3137-454f-8733-9d2e8cccb88f-logs\") pod \"282f4b34-3137-454f-8733-9d2e8cccb88f\" (UID: \"282f4b34-3137-454f-8733-9d2e8cccb88f\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.761073 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282f4b34-3137-454f-8733-9d2e8cccb88f-logs" (OuterVolumeSpecName: "logs") pod "282f4b34-3137-454f-8733-9d2e8cccb88f" (UID: "282f4b34-3137-454f-8733-9d2e8cccb88f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.761229 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/282f4b34-3137-454f-8733-9d2e8cccb88f-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.764860 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/282f4b34-3137-454f-8733-9d2e8cccb88f-kube-api-access-z7jlj" (OuterVolumeSpecName: "kube-api-access-z7jlj") pod "282f4b34-3137-454f-8733-9d2e8cccb88f" (UID: "282f4b34-3137-454f-8733-9d2e8cccb88f"). InnerVolumeSpecName "kube-api-access-z7jlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.766159 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2025e74-2890-43c4-a656-e0f9eb2db5dd","Type":"ContainerDied","Data":"59f4e6d0913b27b26b60bedea31b9ba31a18874ad4f755b67010574ffc17594c"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.766218 4843 scope.go:117] "RemoveContainer" containerID="d6052697ce12d91ea105bb817e1e3fb7b2ae5b3d676ee22525db6d67d198ff19" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.766393 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.767733 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.771866 4843 generic.go:334] "Generic (PLEG): container finished" podID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" exitCode=0 Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.771897 4843 generic.go:334] "Generic (PLEG): container finished" podID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" exitCode=2 Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.771911 4843 generic.go:334] "Generic (PLEG): container finished" podID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" exitCode=0 Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.771922 4843 generic.go:334] "Generic (PLEG): container finished" podID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" exitCode=0 Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.771965 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerDied","Data":"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.771995 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerDied","Data":"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.772008 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerDied","Data":"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.772021 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerDied","Data":"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.772031 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"53c8bb85-a064-4ca4-a541-79be614e88bd","Type":"ContainerDied","Data":"fa12dbfd7723f88e4098104e819d248115c99a107bbfa0fb712d9d813a0878aa"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.773415 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" event={"ID":"286e0246-8d7d-47e9-a44b-3d7acb0ec046","Type":"ContainerDied","Data":"007a0e940fd5b2c7311f1ccae7729f741387bc6430978ebae45660051d1b8115"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.773445 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="007a0e940fd5b2c7311f1ccae7729f741387bc6430978ebae45660051d1b8115" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.773498 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher25dc-account-delete-cr72d" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.787788 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "282f4b34-3137-454f-8733-9d2e8cccb88f" (UID: "282f4b34-3137-454f-8733-9d2e8cccb88f"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.787810 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "282f4b34-3137-454f-8733-9d2e8cccb88f" (UID: "282f4b34-3137-454f-8733-9d2e8cccb88f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.788294 4843 generic.go:334] "Generic (PLEG): container finished" podID="282f4b34-3137-454f-8733-9d2e8cccb88f" containerID="047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff" exitCode=0 Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.788332 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"282f4b34-3137-454f-8733-9d2e8cccb88f","Type":"ContainerDied","Data":"047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.788360 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"282f4b34-3137-454f-8733-9d2e8cccb88f","Type":"ContainerDied","Data":"96ffae1133e1a016b06748d068f9fb14e367f9ad7556dac744dc8983fd1585b1"} Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.788388 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.801635 4843 scope.go:117] "RemoveContainer" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.824602 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.824885 4843 scope.go:117] "RemoveContainer" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.827449 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-config-data" (OuterVolumeSpecName: "config-data") pod "282f4b34-3137-454f-8733-9d2e8cccb88f" (UID: "282f4b34-3137-454f-8733-9d2e8cccb88f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.832437 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.848598 4843 scope.go:117] "RemoveContainer" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.854146 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "282f4b34-3137-454f-8733-9d2e8cccb88f" (UID: "282f4b34-3137-454f-8733-9d2e8cccb88f"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862680 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szz9l\" (UniqueName: \"kubernetes.io/projected/53c8bb85-a064-4ca4-a541-79be614e88bd-kube-api-access-szz9l\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862735 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-log-httpd\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862784 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-run-httpd\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862806 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-config-data\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862849 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-sg-core-conf-yaml\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862868 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-ceilometer-tls-certs\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862901 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-scripts\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.862975 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-combined-ca-bundle\") pod \"53c8bb85-a064-4ca4-a541-79be614e88bd\" (UID: \"53c8bb85-a064-4ca4-a541-79be614e88bd\") " Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.863260 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7jlj\" (UniqueName: \"kubernetes.io/projected/282f4b34-3137-454f-8733-9d2e8cccb88f-kube-api-access-z7jlj\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.863291 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.863300 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.863311 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.863320 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/282f4b34-3137-454f-8733-9d2e8cccb88f-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.864782 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.865160 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.865977 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c8bb85-a064-4ca4-a541-79be614e88bd-kube-api-access-szz9l" (OuterVolumeSpecName: "kube-api-access-szz9l") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "kube-api-access-szz9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.868221 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-scripts" (OuterVolumeSpecName: "scripts") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.877602 4843 scope.go:117] "RemoveContainer" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.887229 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.899865 4843 scope.go:117] "RemoveContainer" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" Mar 14 09:52:39 crc kubenswrapper[4843]: E0314 09:52:39.900202 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": container with ID starting with 1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e not found: ID does not exist" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.900238 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e"} err="failed to get container status \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": rpc error: code = NotFound desc = could not find container \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": container with ID starting with 1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.900260 4843 scope.go:117] "RemoveContainer" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" Mar 14 09:52:39 crc kubenswrapper[4843]: E0314 09:52:39.900544 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": container with ID starting with 2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70 not found: ID does not exist" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.900604 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70"} err="failed to get container status \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": rpc error: code = NotFound desc = could not find container \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": container with ID starting with 2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.900633 4843 scope.go:117] "RemoveContainer" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" Mar 14 09:52:39 crc kubenswrapper[4843]: E0314 09:52:39.901375 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": container with ID starting with 206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b not found: ID does not exist" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.901422 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b"} err="failed to get container status \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": rpc error: code = NotFound desc = could not find container \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": container with ID starting with 206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.901448 4843 scope.go:117] "RemoveContainer" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" Mar 14 09:52:39 crc kubenswrapper[4843]: E0314 09:52:39.901756 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": container with ID starting with 84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983 not found: ID does not exist" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.901779 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983"} err="failed to get container status \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": rpc error: code = NotFound desc = could not find container \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": container with ID starting with 84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.901795 4843 scope.go:117] "RemoveContainer" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.901976 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e"} err="failed to get container status \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": rpc error: code = NotFound desc = could not find container \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": container with ID starting with 1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.901995 4843 scope.go:117] "RemoveContainer" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.902924 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70"} err="failed to get container status \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": rpc error: code = NotFound desc = could not find container \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": container with ID starting with 2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.902964 4843 scope.go:117] "RemoveContainer" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.903328 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b"} err="failed to get container status \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": rpc error: code = NotFound desc = could not find container \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": container with ID starting with 206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.903395 4843 scope.go:117] "RemoveContainer" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.903698 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983"} err="failed to get container status \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": rpc error: code = NotFound desc = could not find container \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": container with ID starting with 84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.903761 4843 scope.go:117] "RemoveContainer" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904122 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e"} err="failed to get container status \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": rpc error: code = NotFound desc = could not find container \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": container with ID starting with 1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904142 4843 scope.go:117] "RemoveContainer" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904390 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70"} err="failed to get container status \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": rpc error: code = NotFound desc = could not find container \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": container with ID starting with 2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904411 4843 scope.go:117] "RemoveContainer" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904627 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b"} err="failed to get container status \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": rpc error: code = NotFound desc = could not find container \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": container with ID starting with 206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904651 4843 scope.go:117] "RemoveContainer" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904850 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983"} err="failed to get container status \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": rpc error: code = NotFound desc = could not find container \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": container with ID starting with 84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.904876 4843 scope.go:117] "RemoveContainer" containerID="1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.905065 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e"} err="failed to get container status \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": rpc error: code = NotFound desc = could not find container \"1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e\": container with ID starting with 1c852121386d15cb93e37a8f5decb0af15771e6bbe70e5a8804318b18022949e not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.905093 4843 scope.go:117] "RemoveContainer" containerID="2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.905331 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70"} err="failed to get container status \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": rpc error: code = NotFound desc = could not find container \"2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70\": container with ID starting with 2b9f24d6af563c1f8dfe02169cab2673ff2f0b013c653e2582a5ad2363487d70 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.905352 4843 scope.go:117] "RemoveContainer" containerID="206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.906168 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b"} err="failed to get container status \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": rpc error: code = NotFound desc = could not find container \"206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b\": container with ID starting with 206bafb3bb9b24116928838b6d982cd0fc1fd37f1970e9303adb302d124b6a1b not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.906195 4843 scope.go:117] "RemoveContainer" containerID="84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.906608 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983"} err="failed to get container status \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": rpc error: code = NotFound desc = could not find container \"84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983\": container with ID starting with 84884203f5e883b3634142ef82bc347b56f45189751d219dc31ff573480f0983 not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.906630 4843 scope.go:117] "RemoveContainer" containerID="047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.921704 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.928065 4843 scope.go:117] "RemoveContainer" containerID="047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff" Mar 14 09:52:39 crc kubenswrapper[4843]: E0314 09:52:39.928426 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff\": container with ID starting with 047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff not found: ID does not exist" containerID="047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.928485 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff"} err="failed to get container status \"047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff\": rpc error: code = NotFound desc = could not find container \"047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff\": container with ID starting with 047324b48d02d9bd2d572ba8dca94484342ecfe00e6680ea49ae1859dcdf2dff not found: ID does not exist" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.931864 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.954264 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-config-data" (OuterVolumeSpecName: "config-data") pod "53c8bb85-a064-4ca4-a541-79be614e88bd" (UID: "53c8bb85-a064-4ca4-a541-79be614e88bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966026 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966055 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szz9l\" (UniqueName: \"kubernetes.io/projected/53c8bb85-a064-4ca4-a541-79be614e88bd-kube-api-access-szz9l\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966065 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966073 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53c8bb85-a064-4ca4-a541-79be614e88bd-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966084 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966092 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966100 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:39 crc kubenswrapper[4843]: I0314 09:52:39.966107 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8bb85-a064-4ca4-a541-79be614e88bd-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.124877 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.131917 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.808072 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.843734 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher25dc-account-delete-cr72d"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.854256 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-25dc-account-create-update-qssln"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.864733 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher25dc-account-delete-cr72d"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.874044 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-fzhnw"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.882483 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-25dc-account-create-update-qssln"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.888713 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-fzhnw"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.897670 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.903491 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912264 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912663 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-kuttl-api-log" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912685 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-kuttl-api-log" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912726 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286e0246-8d7d-47e9-a44b-3d7acb0ec046" containerName="mariadb-account-delete" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912733 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="286e0246-8d7d-47e9-a44b-3d7acb0ec046" containerName="mariadb-account-delete" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912740 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-central-agent" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912747 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-central-agent" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912757 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-api" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912762 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-api" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912773 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="proxy-httpd" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912779 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="proxy-httpd" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912791 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" containerName="watcher-applier" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912796 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" containerName="watcher-applier" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912810 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-notification-agent" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912816 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-notification-agent" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912834 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282f4b34-3137-454f-8733-9d2e8cccb88f" containerName="watcher-decision-engine" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912841 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="282f4b34-3137-454f-8733-9d2e8cccb88f" containerName="watcher-decision-engine" Mar 14 09:52:40 crc kubenswrapper[4843]: E0314 09:52:40.912851 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="sg-core" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.912856 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="sg-core" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913001 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-central-agent" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913009 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-kuttl-api-log" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913020 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="proxy-httpd" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913027 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="286e0246-8d7d-47e9-a44b-3d7acb0ec046" containerName="mariadb-account-delete" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913037 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="ceilometer-notification-agent" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913046 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="282f4b34-3137-454f-8733-9d2e8cccb88f" containerName="watcher-decision-engine" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913057 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" containerName="sg-core" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913067 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" containerName="watcher-applier" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.913075 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d11aac1-e789-4561-b182-1e6f150f10ac" containerName="watcher-api" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.918305 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.927752 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.928062 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.928316 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.942401 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-qwqwz"] Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.954010 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.978917 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-config-data\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.978982 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.979035 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-run-httpd\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.979069 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.979091 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-scripts\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.979164 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.979199 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-log-httpd\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.979221 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfqk5\" (UniqueName: \"kubernetes.io/projected/5a623a35-5d11-45c3-8874-8993915bb09e-kube-api-access-hfqk5\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:40 crc kubenswrapper[4843]: I0314 09:52:40.982811 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.013709 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-qwqwz"] Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.034306 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-gd7kp"] Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.035368 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.039348 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.043862 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-gd7kp"] Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.080598 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hjzz\" (UniqueName: \"kubernetes.io/projected/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-kube-api-access-8hjzz\") pod \"watcher-db-create-qwqwz\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.080648 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.080696 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-run-httpd\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.080714 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4pvj\" (UniqueName: \"kubernetes.io/projected/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-kube-api-access-x4pvj\") pod \"watcher-test-account-create-update-gd7kp\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.080850 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.080902 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-scripts\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081028 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-operator-scripts\") pod \"watcher-db-create-qwqwz\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081097 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081151 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-log-httpd\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081158 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-run-httpd\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081176 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfqk5\" (UniqueName: \"kubernetes.io/projected/5a623a35-5d11-45c3-8874-8993915bb09e-kube-api-access-hfqk5\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081218 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-operator-scripts\") pod \"watcher-test-account-create-update-gd7kp\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.081256 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-config-data\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.082056 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-log-httpd\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.086795 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.086822 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.110850 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.112810 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-config-data\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.115028 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfqk5\" (UniqueName: \"kubernetes.io/projected/5a623a35-5d11-45c3-8874-8993915bb09e-kube-api-access-hfqk5\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.116833 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-scripts\") pod \"ceilometer-0\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.182500 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-operator-scripts\") pod \"watcher-db-create-qwqwz\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.182594 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-operator-scripts\") pod \"watcher-test-account-create-update-gd7kp\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.182642 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hjzz\" (UniqueName: \"kubernetes.io/projected/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-kube-api-access-8hjzz\") pod \"watcher-db-create-qwqwz\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.182702 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4pvj\" (UniqueName: \"kubernetes.io/projected/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-kube-api-access-x4pvj\") pod \"watcher-test-account-create-update-gd7kp\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.183458 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-operator-scripts\") pod \"watcher-test-account-create-update-gd7kp\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.183458 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-operator-scripts\") pod \"watcher-db-create-qwqwz\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.198214 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4pvj\" (UniqueName: \"kubernetes.io/projected/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-kube-api-access-x4pvj\") pod \"watcher-test-account-create-update-gd7kp\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.198401 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hjzz\" (UniqueName: \"kubernetes.io/projected/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-kube-api-access-8hjzz\") pod \"watcher-db-create-qwqwz\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.269359 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.304349 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.365674 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.370536 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="282f4b34-3137-454f-8733-9d2e8cccb88f" path="/var/lib/kubelet/pods/282f4b34-3137-454f-8733-9d2e8cccb88f/volumes" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.371211 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="286e0246-8d7d-47e9-a44b-3d7acb0ec046" path="/var/lib/kubelet/pods/286e0246-8d7d-47e9-a44b-3d7acb0ec046/volumes" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.371725 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53c8bb85-a064-4ca4-a541-79be614e88bd" path="/var/lib/kubelet/pods/53c8bb85-a064-4ca4-a541-79be614e88bd/volumes" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.375753 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b" path="/var/lib/kubelet/pods/af6b7c42-02ea-4df8-a95f-9a2b3ae2e16b/volumes" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.376375 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2025e74-2890-43c4-a656-e0f9eb2db5dd" path="/var/lib/kubelet/pods/b2025e74-2890-43c4-a656-e0f9eb2db5dd/volumes" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.376927 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef4afd5-c0a9-4443-aa51-358adb5b9014" path="/var/lib/kubelet/pods/bef4afd5-c0a9-4443-aa51-358adb5b9014/volumes" Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.804977 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:52:41 crc kubenswrapper[4843]: I0314 09:52:41.817172 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerStarted","Data":"0b9e9fdb29deb589536828aff86ba957937638b6e7b15dc7c3cd72abc0d811c8"} Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.005750 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-gd7kp"] Mar 14 09:52:42 crc kubenswrapper[4843]: W0314 09:52:42.012494 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8ab63c1_c44d_4957_bdc7_5b3fd7e92795.slice/crio-761f9e465de6015649b3119fd3e9502cf089cd58f8c28c768bec32b19c79f582 WatchSource:0}: Error finding container 761f9e465de6015649b3119fd3e9502cf089cd58f8c28c768bec32b19c79f582: Status 404 returned error can't find the container with id 761f9e465de6015649b3119fd3e9502cf089cd58f8c28c768bec32b19c79f582 Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.012674 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-qwqwz"] Mar 14 09:52:42 crc kubenswrapper[4843]: W0314 09:52:42.013705 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d3682d7_b738_4ca1_9a9a_10b4d5fde44a.slice/crio-12c1376055076b4171721e7b43bb854256e642408903e6d8001b543fb9e73c9f WatchSource:0}: Error finding container 12c1376055076b4171721e7b43bb854256e642408903e6d8001b543fb9e73c9f: Status 404 returned error can't find the container with id 12c1376055076b4171721e7b43bb854256e642408903e6d8001b543fb9e73c9f Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.826888 4843 generic.go:334] "Generic (PLEG): container finished" podID="c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" containerID="c7d3d78ca542af44a37d98eda1a591d5334e6a5633feff7428bd668f99e0354b" exitCode=0 Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.826998 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-qwqwz" event={"ID":"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795","Type":"ContainerDied","Data":"c7d3d78ca542af44a37d98eda1a591d5334e6a5633feff7428bd668f99e0354b"} Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.827589 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-qwqwz" event={"ID":"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795","Type":"ContainerStarted","Data":"761f9e465de6015649b3119fd3e9502cf089cd58f8c28c768bec32b19c79f582"} Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.830453 4843 generic.go:334] "Generic (PLEG): container finished" podID="8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" containerID="937a6ad5ea002abf3145c18da1261d4469f21d3b2468754ae71719789c6ed630" exitCode=0 Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.830564 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" event={"ID":"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a","Type":"ContainerDied","Data":"937a6ad5ea002abf3145c18da1261d4469f21d3b2468754ae71719789c6ed630"} Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.830614 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" event={"ID":"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a","Type":"ContainerStarted","Data":"12c1376055076b4171721e7b43bb854256e642408903e6d8001b543fb9e73c9f"} Mar 14 09:52:42 crc kubenswrapper[4843]: I0314 09:52:42.837316 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerStarted","Data":"0c3b43f370bf072d24328cff4e51458308206a6e8e5eabae7a4b17728839746b"} Mar 14 09:52:43 crc kubenswrapper[4843]: I0314 09:52:43.361335 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:52:43 crc kubenswrapper[4843]: E0314 09:52:43.361703 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:52:43 crc kubenswrapper[4843]: I0314 09:52:43.855467 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerStarted","Data":"4e0d55d20991c5006332a308ead2ac66aac4ab0e0541dc11565c005fd1f2704b"} Mar 14 09:52:43 crc kubenswrapper[4843]: I0314 09:52:43.855757 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerStarted","Data":"8b857a047d7a307b40f75b574496934e8088d289b855493840a40cb1d916287f"} Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.308379 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.345848 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-operator-scripts\") pod \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.346017 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4pvj\" (UniqueName: \"kubernetes.io/projected/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-kube-api-access-x4pvj\") pod \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\" (UID: \"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a\") " Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.346680 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" (UID: "8d3682d7-b738-4ca1-9a9a-10b4d5fde44a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.376369 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-kube-api-access-x4pvj" (OuterVolumeSpecName: "kube-api-access-x4pvj") pod "8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" (UID: "8d3682d7-b738-4ca1-9a9a-10b4d5fde44a"). InnerVolumeSpecName "kube-api-access-x4pvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.408636 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.447747 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hjzz\" (UniqueName: \"kubernetes.io/projected/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-kube-api-access-8hjzz\") pod \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.447812 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-operator-scripts\") pod \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\" (UID: \"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795\") " Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.448231 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.448253 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4pvj\" (UniqueName: \"kubernetes.io/projected/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a-kube-api-access-x4pvj\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.448361 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" (UID: "c8ab63c1-c44d-4957-bdc7-5b3fd7e92795"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.465384 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-kube-api-access-8hjzz" (OuterVolumeSpecName: "kube-api-access-8hjzz") pod "c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" (UID: "c8ab63c1-c44d-4957-bdc7-5b3fd7e92795"). InnerVolumeSpecName "kube-api-access-8hjzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.549577 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hjzz\" (UniqueName: \"kubernetes.io/projected/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-kube-api-access-8hjzz\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.549629 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.864727 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-qwqwz" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.864720 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-qwqwz" event={"ID":"c8ab63c1-c44d-4957-bdc7-5b3fd7e92795","Type":"ContainerDied","Data":"761f9e465de6015649b3119fd3e9502cf089cd58f8c28c768bec32b19c79f582"} Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.864864 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="761f9e465de6015649b3119fd3e9502cf089cd58f8c28c768bec32b19c79f582" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.865935 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" event={"ID":"8d3682d7-b738-4ca1-9a9a-10b4d5fde44a","Type":"ContainerDied","Data":"12c1376055076b4171721e7b43bb854256e642408903e6d8001b543fb9e73c9f"} Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.865949 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12c1376055076b4171721e7b43bb854256e642408903e6d8001b543fb9e73c9f" Mar 14 09:52:44 crc kubenswrapper[4843]: I0314 09:52:44.866018 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-test-account-create-update-gd7kp" Mar 14 09:52:45 crc kubenswrapper[4843]: I0314 09:52:45.874984 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerStarted","Data":"75fa0c663ae9908d5b20a2bcbcc935c04d1a6446e6cd49da066c6c5a63421cb9"} Mar 14 09:52:45 crc kubenswrapper[4843]: I0314 09:52:45.875858 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:52:45 crc kubenswrapper[4843]: I0314 09:52:45.904628 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.561555503 podStartE2EDuration="5.904608257s" podCreationTimestamp="2026-03-14 09:52:40 +0000 UTC" firstStartedPulling="2026-03-14 09:52:41.811010834 +0000 UTC m=+2529.123621962" lastFinishedPulling="2026-03-14 09:52:45.154063588 +0000 UTC m=+2532.466674716" observedRunningTime="2026-03-14 09:52:45.901740657 +0000 UTC m=+2533.214351825" watchObservedRunningTime="2026-03-14 09:52:45.904608257 +0000 UTC m=+2533.217219395" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.299078 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-br9mg"] Mar 14 09:52:46 crc kubenswrapper[4843]: E0314 09:52:46.299497 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" containerName="mariadb-database-create" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.299518 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" containerName="mariadb-database-create" Mar 14 09:52:46 crc kubenswrapper[4843]: E0314 09:52:46.299561 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" containerName="mariadb-account-create-update" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.299572 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" containerName="mariadb-account-create-update" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.299783 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" containerName="mariadb-account-create-update" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.299822 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" containerName="mariadb-database-create" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.300427 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.302709 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.302756 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-992jk" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.329429 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-br9mg"] Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.379186 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f88bm\" (UniqueName: \"kubernetes.io/projected/238d64c0-54b7-4541-a7f0-01a95fc2a821-kube-api-access-f88bm\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.379503 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-config-data\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.379693 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-db-sync-config-data\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.379896 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.481440 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f88bm\" (UniqueName: \"kubernetes.io/projected/238d64c0-54b7-4541-a7f0-01a95fc2a821-kube-api-access-f88bm\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.481688 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-config-data\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.481783 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-db-sync-config-data\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.481902 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.495212 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-config-data\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.495368 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.495395 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-db-sync-config-data\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.497164 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f88bm\" (UniqueName: \"kubernetes.io/projected/238d64c0-54b7-4541-a7f0-01a95fc2a821-kube-api-access-f88bm\") pod \"watcher-kuttl-db-sync-br9mg\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:46 crc kubenswrapper[4843]: I0314 09:52:46.617864 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:47 crc kubenswrapper[4843]: I0314 09:52:47.086028 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-br9mg"] Mar 14 09:52:47 crc kubenswrapper[4843]: I0314 09:52:47.939823 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" event={"ID":"238d64c0-54b7-4541-a7f0-01a95fc2a821","Type":"ContainerStarted","Data":"df704e57298033b22df6bfb1642aec244e6d65c286f5e7e9971184d849a3cf49"} Mar 14 09:52:47 crc kubenswrapper[4843]: I0314 09:52:47.940628 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" event={"ID":"238d64c0-54b7-4541-a7f0-01a95fc2a821","Type":"ContainerStarted","Data":"e5b40425abfd8349ebc7ea44e4286e490b6a6425fb54f0b6105f2b68566325a2"} Mar 14 09:52:47 crc kubenswrapper[4843]: I0314 09:52:47.958726 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" podStartSLOduration=1.9587063470000001 podStartE2EDuration="1.958706347s" podCreationTimestamp="2026-03-14 09:52:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:47.956702168 +0000 UTC m=+2535.269313296" watchObservedRunningTime="2026-03-14 09:52:47.958706347 +0000 UTC m=+2535.271317485" Mar 14 09:52:49 crc kubenswrapper[4843]: I0314 09:52:49.960586 4843 generic.go:334] "Generic (PLEG): container finished" podID="238d64c0-54b7-4541-a7f0-01a95fc2a821" containerID="df704e57298033b22df6bfb1642aec244e6d65c286f5e7e9971184d849a3cf49" exitCode=0 Mar 14 09:52:49 crc kubenswrapper[4843]: I0314 09:52:49.960823 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" event={"ID":"238d64c0-54b7-4541-a7f0-01a95fc2a821","Type":"ContainerDied","Data":"df704e57298033b22df6bfb1642aec244e6d65c286f5e7e9971184d849a3cf49"} Mar 14 09:52:50 crc kubenswrapper[4843]: I0314 09:52:50.551909 4843 scope.go:117] "RemoveContainer" containerID="2f1821f65759fdd2e3f3091ac7d5acfe8befd75222d5455e64a6d01613041443" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.384973 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.471767 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-combined-ca-bundle\") pod \"238d64c0-54b7-4541-a7f0-01a95fc2a821\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.472052 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-db-sync-config-data\") pod \"238d64c0-54b7-4541-a7f0-01a95fc2a821\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.472076 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f88bm\" (UniqueName: \"kubernetes.io/projected/238d64c0-54b7-4541-a7f0-01a95fc2a821-kube-api-access-f88bm\") pod \"238d64c0-54b7-4541-a7f0-01a95fc2a821\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.472098 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-config-data\") pod \"238d64c0-54b7-4541-a7f0-01a95fc2a821\" (UID: \"238d64c0-54b7-4541-a7f0-01a95fc2a821\") " Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.477467 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "238d64c0-54b7-4541-a7f0-01a95fc2a821" (UID: "238d64c0-54b7-4541-a7f0-01a95fc2a821"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.478200 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/238d64c0-54b7-4541-a7f0-01a95fc2a821-kube-api-access-f88bm" (OuterVolumeSpecName: "kube-api-access-f88bm") pod "238d64c0-54b7-4541-a7f0-01a95fc2a821" (UID: "238d64c0-54b7-4541-a7f0-01a95fc2a821"). InnerVolumeSpecName "kube-api-access-f88bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.495518 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "238d64c0-54b7-4541-a7f0-01a95fc2a821" (UID: "238d64c0-54b7-4541-a7f0-01a95fc2a821"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.528414 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-config-data" (OuterVolumeSpecName: "config-data") pod "238d64c0-54b7-4541-a7f0-01a95fc2a821" (UID: "238d64c0-54b7-4541-a7f0-01a95fc2a821"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.574463 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.574709 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.574775 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f88bm\" (UniqueName: \"kubernetes.io/projected/238d64c0-54b7-4541-a7f0-01a95fc2a821-kube-api-access-f88bm\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.574841 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/238d64c0-54b7-4541-a7f0-01a95fc2a821-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.981885 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" event={"ID":"238d64c0-54b7-4541-a7f0-01a95fc2a821","Type":"ContainerDied","Data":"e5b40425abfd8349ebc7ea44e4286e490b6a6425fb54f0b6105f2b68566325a2"} Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.981942 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b40425abfd8349ebc7ea44e4286e490b6a6425fb54f0b6105f2b68566325a2" Mar 14 09:52:51 crc kubenswrapper[4843]: I0314 09:52:51.981969 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-br9mg" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.320441 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:52 crc kubenswrapper[4843]: E0314 09:52:52.320802 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238d64c0-54b7-4541-a7f0-01a95fc2a821" containerName="watcher-kuttl-db-sync" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.320817 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="238d64c0-54b7-4541-a7f0-01a95fc2a821" containerName="watcher-kuttl-db-sync" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.320959 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="238d64c0-54b7-4541-a7f0-01a95fc2a821" containerName="watcher-kuttl-db-sync" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.321931 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.324452 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.324858 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-992jk" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.341735 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.364285 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.365577 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.389291 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.390070 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqr4v\" (UniqueName: \"kubernetes.io/projected/df4de54e-5485-44e1-80bc-46f18428bb9d-kube-api-access-vqr4v\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.390306 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.390423 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.390571 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.390717 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/447dff52-cadc-414d-9ba3-73d307d1bb35-logs\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.390809 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.391055 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df4de54e-5485-44e1-80bc-46f18428bb9d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.391101 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.391139 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qcc7\" (UniqueName: \"kubernetes.io/projected/447dff52-cadc-414d-9ba3-73d307d1bb35-kube-api-access-7qcc7\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.391155 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.391238 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.409808 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.442870 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.443909 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.449337 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.469174 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.493743 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.493809 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0f0536-eea6-4e03-86a4-f3a077128d0e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.493837 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqr4v\" (UniqueName: \"kubernetes.io/projected/df4de54e-5485-44e1-80bc-46f18428bb9d-kube-api-access-vqr4v\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.493873 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.493900 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.493999 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494051 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/447dff52-cadc-414d-9ba3-73d307d1bb35-logs\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494070 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494123 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494243 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df4de54e-5485-44e1-80bc-46f18428bb9d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494298 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494322 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg4q6\" (UniqueName: \"kubernetes.io/projected/6a0f0536-eea6-4e03-86a4-f3a077128d0e-kube-api-access-sg4q6\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494351 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494378 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qcc7\" (UniqueName: \"kubernetes.io/projected/447dff52-cadc-414d-9ba3-73d307d1bb35-kube-api-access-7qcc7\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494399 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494420 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494505 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/447dff52-cadc-414d-9ba3-73d307d1bb35-logs\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.494515 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.498188 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.500673 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df4de54e-5485-44e1-80bc-46f18428bb9d-logs\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.501614 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.501992 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.504527 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.509770 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.509842 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.509906 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.518370 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qcc7\" (UniqueName: \"kubernetes.io/projected/447dff52-cadc-414d-9ba3-73d307d1bb35-kube-api-access-7qcc7\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.519334 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqr4v\" (UniqueName: \"kubernetes.io/projected/df4de54e-5485-44e1-80bc-46f18428bb9d-kube-api-access-vqr4v\") pod \"watcher-kuttl-api-0\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.523803 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.533444 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.534520 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.544956 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.548184 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.595907 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp2j2\" (UniqueName: \"kubernetes.io/projected/6f569cfa-dba2-4b00-8dd9-052336b7014a-kube-api-access-dp2j2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.595944 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.595991 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596014 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0f0536-eea6-4e03-86a4-f3a077128d0e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596127 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596211 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596259 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg4q6\" (UniqueName: \"kubernetes.io/projected/6a0f0536-eea6-4e03-86a4-f3a077128d0e-kube-api-access-sg4q6\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596301 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596334 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596362 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596416 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f569cfa-dba2-4b00-8dd9-052336b7014a-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.596802 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0f0536-eea6-4e03-86a4-f3a077128d0e-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.600325 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.600788 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.600919 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.612872 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg4q6\" (UniqueName: \"kubernetes.io/projected/6a0f0536-eea6-4e03-86a4-f3a077128d0e-kube-api-access-sg4q6\") pod \"watcher-kuttl-applier-0\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.635895 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.698008 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.698654 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f569cfa-dba2-4b00-8dd9-052336b7014a-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.698908 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp2j2\" (UniqueName: \"kubernetes.io/projected/6f569cfa-dba2-4b00-8dd9-052336b7014a-kube-api-access-dp2j2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.699003 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.699109 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.699252 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.699383 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.699523 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f569cfa-dba2-4b00-8dd9-052336b7014a-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.703084 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.703467 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.704911 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.705666 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.724286 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp2j2\" (UniqueName: \"kubernetes.io/projected/6f569cfa-dba2-4b00-8dd9-052336b7014a-kube-api-access-dp2j2\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.771678 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:52:52 crc kubenswrapper[4843]: I0314 09:52:52.896653 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:52:53 crc kubenswrapper[4843]: I0314 09:52:53.095969 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:52:53 crc kubenswrapper[4843]: I0314 09:52:53.197452 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:52:53 crc kubenswrapper[4843]: W0314 09:52:53.200989 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod447dff52_cadc_414d_9ba3_73d307d1bb35.slice/crio-294d930deefc60828f0e690a338446f67d0dbebb3f314be13400c969f4def89b WatchSource:0}: Error finding container 294d930deefc60828f0e690a338446f67d0dbebb3f314be13400c969f4def89b: Status 404 returned error can't find the container with id 294d930deefc60828f0e690a338446f67d0dbebb3f314be13400c969f4def89b Mar 14 09:52:53 crc kubenswrapper[4843]: I0314 09:52:53.311033 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:52:53 crc kubenswrapper[4843]: W0314 09:52:53.315836 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a0f0536_eea6_4e03_86a4_f3a077128d0e.slice/crio-14b34d60805f1b2b9cdb48622e6b79674f24f9aa6a091aa2a1fc0e5c4a54a86c WatchSource:0}: Error finding container 14b34d60805f1b2b9cdb48622e6b79674f24f9aa6a091aa2a1fc0e5c4a54a86c: Status 404 returned error can't find the container with id 14b34d60805f1b2b9cdb48622e6b79674f24f9aa6a091aa2a1fc0e5c4a54a86c Mar 14 09:52:53 crc kubenswrapper[4843]: I0314 09:52:53.504341 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.005160 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"447dff52-cadc-414d-9ba3-73d307d1bb35","Type":"ContainerStarted","Data":"518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.005222 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"447dff52-cadc-414d-9ba3-73d307d1bb35","Type":"ContainerStarted","Data":"924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.005242 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"447dff52-cadc-414d-9ba3-73d307d1bb35","Type":"ContainerStarted","Data":"294d930deefc60828f0e690a338446f67d0dbebb3f314be13400c969f4def89b"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.005421 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.006700 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"df4de54e-5485-44e1-80bc-46f18428bb9d","Type":"ContainerStarted","Data":"f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.006751 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"df4de54e-5485-44e1-80bc-46f18428bb9d","Type":"ContainerStarted","Data":"28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.006763 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"df4de54e-5485-44e1-80bc-46f18428bb9d","Type":"ContainerStarted","Data":"5f5888550cab36d7b3fcd903ffc22d261a4276273923ad3f2435e9243f9db3f7"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.007260 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.008365 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"6f569cfa-dba2-4b00-8dd9-052336b7014a","Type":"ContainerStarted","Data":"019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.008394 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"6f569cfa-dba2-4b00-8dd9-052336b7014a","Type":"ContainerStarted","Data":"b2c6df07f138bf153032f1534b356e5512ec4b661368f2c5a888bc78550efb6e"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.010695 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6a0f0536-eea6-4e03-86a4-f3a077128d0e","Type":"ContainerStarted","Data":"e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.010754 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6a0f0536-eea6-4e03-86a4-f3a077128d0e","Type":"ContainerStarted","Data":"14b34d60805f1b2b9cdb48622e6b79674f24f9aa6a091aa2a1fc0e5c4a54a86c"} Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.044604 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-1" podStartSLOduration=2.044584256 podStartE2EDuration="2.044584256s" podCreationTimestamp="2026-03-14 09:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:54.042859303 +0000 UTC m=+2541.355470441" watchObservedRunningTime="2026-03-14 09:52:54.044584256 +0000 UTC m=+2541.357195384" Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.101211 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.101189462 podStartE2EDuration="2.101189462s" podCreationTimestamp="2026-03-14 09:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:54.07660232 +0000 UTC m=+2541.389213468" watchObservedRunningTime="2026-03-14 09:52:54.101189462 +0000 UTC m=+2541.413800610" Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.122813 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.122795342 podStartE2EDuration="2.122795342s" podCreationTimestamp="2026-03-14 09:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:54.099211524 +0000 UTC m=+2541.411822662" watchObservedRunningTime="2026-03-14 09:52:54.122795342 +0000 UTC m=+2541.435406470" Mar 14 09:52:54 crc kubenswrapper[4843]: I0314 09:52:54.125915 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.125902068 podStartE2EDuration="2.125902068s" podCreationTimestamp="2026-03-14 09:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:52:54.118075136 +0000 UTC m=+2541.430686284" watchObservedRunningTime="2026-03-14 09:52:54.125902068 +0000 UTC m=+2541.438513196" Mar 14 09:52:54 crc kubenswrapper[4843]: E0314 09:52:54.617458 4843 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.162:51564->38.102.83.162:34757: write tcp 38.102.83.162:51564->38.102.83.162:34757: write: connection reset by peer Mar 14 09:52:56 crc kubenswrapper[4843]: I0314 09:52:56.229551 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:56 crc kubenswrapper[4843]: I0314 09:52:56.340009 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:52:56 crc kubenswrapper[4843]: I0314 09:52:56.763074 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:57 crc kubenswrapper[4843]: I0314 09:52:57.035123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"28e8d32901c9974c782e1e9aadfeb56ef673f0cf23ac2a95a7d8e3c463d945a9"} Mar 14 09:52:57 crc kubenswrapper[4843]: I0314 09:52:57.635991 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:52:57 crc kubenswrapper[4843]: I0314 09:52:57.698320 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:52:57 crc kubenswrapper[4843]: I0314 09:52:57.772331 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.136166 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph"] Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.137651 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.139774 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.141147 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-scripts" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.160035 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph"] Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.312614 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-config-data\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.312676 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-scripts-volume\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.312806 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-combined-ca-bundle\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.312834 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfnrm\" (UniqueName: \"kubernetes.io/projected/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-kube-api-access-rfnrm\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.414608 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-combined-ca-bundle\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.414676 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfnrm\" (UniqueName: \"kubernetes.io/projected/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-kube-api-access-rfnrm\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.414749 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-config-data\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.414794 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-scripts-volume\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.420439 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-scripts-volume\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.420846 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-config-data\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.421541 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-combined-ca-bundle\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.436608 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfnrm\" (UniqueName: \"kubernetes.io/projected/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-kube-api-access-rfnrm\") pod \"watcher-kuttl-db-purge-29558033-fhmph\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.461963 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:00 crc kubenswrapper[4843]: I0314 09:53:00.956944 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph"] Mar 14 09:53:00 crc kubenswrapper[4843]: W0314 09:53:00.957315 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e91ce4b_a255_4a4b_ad9f_0510a8db94b1.slice/crio-0b9d7929977db6977ed45ea33dfccc12c6cb2421b368a0f818dea66eda48be3d WatchSource:0}: Error finding container 0b9d7929977db6977ed45ea33dfccc12c6cb2421b368a0f818dea66eda48be3d: Status 404 returned error can't find the container with id 0b9d7929977db6977ed45ea33dfccc12c6cb2421b368a0f818dea66eda48be3d Mar 14 09:53:01 crc kubenswrapper[4843]: I0314 09:53:01.134807 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" event={"ID":"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1","Type":"ContainerStarted","Data":"0b9d7929977db6977ed45ea33dfccc12c6cb2421b368a0f818dea66eda48be3d"} Mar 14 09:53:01 crc kubenswrapper[4843]: I0314 09:53:01.156300 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" podStartSLOduration=1.156267465 podStartE2EDuration="1.156267465s" podCreationTimestamp="2026-03-14 09:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:01.150906553 +0000 UTC m=+2548.463517761" watchObservedRunningTime="2026-03-14 09:53:01.156267465 +0000 UTC m=+2548.468878603" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.144237 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" event={"ID":"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1","Type":"ContainerStarted","Data":"87a3b4b0238d56ed494a207a396ee7efb484b018287a60013c412c82c98f0172"} Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.636299 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.649949 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.698918 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.711071 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.773144 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.830168 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.897111 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:02 crc kubenswrapper[4843]: I0314 09:53:02.936156 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:03 crc kubenswrapper[4843]: I0314 09:53:03.152304 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:03 crc kubenswrapper[4843]: I0314 09:53:03.158907 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:03 crc kubenswrapper[4843]: I0314 09:53:03.159543 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:53:03 crc kubenswrapper[4843]: I0314 09:53:03.187961 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:03 crc kubenswrapper[4843]: I0314 09:53:03.189939 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:04 crc kubenswrapper[4843]: I0314 09:53:04.159564 4843 generic.go:334] "Generic (PLEG): container finished" podID="2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" containerID="87a3b4b0238d56ed494a207a396ee7efb484b018287a60013c412c82c98f0172" exitCode=0 Mar 14 09:53:04 crc kubenswrapper[4843]: I0314 09:53:04.159731 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" event={"ID":"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1","Type":"ContainerDied","Data":"87a3b4b0238d56ed494a207a396ee7efb484b018287a60013c412c82c98f0172"} Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.537417 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.559655 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.560028 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-central-agent" containerID="cri-o://0c3b43f370bf072d24328cff4e51458308206a6e8e5eabae7a4b17728839746b" gracePeriod=30 Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.560131 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="sg-core" containerID="cri-o://4e0d55d20991c5006332a308ead2ac66aac4ab0e0541dc11565c005fd1f2704b" gracePeriod=30 Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.560172 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-notification-agent" containerID="cri-o://8b857a047d7a307b40f75b574496934e8088d289b855493840a40cb1d916287f" gracePeriod=30 Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.560295 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="proxy-httpd" containerID="cri-o://75fa0c663ae9908d5b20a2bcbcc935c04d1a6446e6cd49da066c6c5a63421cb9" gracePeriod=30 Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.582758 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.226:3000/\": EOF" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.609955 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfnrm\" (UniqueName: \"kubernetes.io/projected/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-kube-api-access-rfnrm\") pod \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.610179 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-combined-ca-bundle\") pod \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.610245 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-config-data\") pod \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.610300 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-scripts-volume\") pod \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\" (UID: \"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1\") " Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.616666 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-scripts-volume" (OuterVolumeSpecName: "scripts-volume") pod "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" (UID: "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1"). InnerVolumeSpecName "scripts-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.616834 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-kube-api-access-rfnrm" (OuterVolumeSpecName: "kube-api-access-rfnrm") pod "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" (UID: "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1"). InnerVolumeSpecName "kube-api-access-rfnrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.638824 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" (UID: "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.659400 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-config-data" (OuterVolumeSpecName: "config-data") pod "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" (UID: "2e91ce4b-a255-4a4b-ad9f-0510a8db94b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.712038 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.712075 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.712083 4843 reconciler_common.go:293] "Volume detached for volume \"scripts-volume\" (UniqueName: \"kubernetes.io/secret/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-scripts-volume\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:05 crc kubenswrapper[4843]: I0314 09:53:05.712093 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfnrm\" (UniqueName: \"kubernetes.io/projected/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1-kube-api-access-rfnrm\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.190012 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" event={"ID":"2e91ce4b-a255-4a4b-ad9f-0510a8db94b1","Type":"ContainerDied","Data":"0b9d7929977db6977ed45ea33dfccc12c6cb2421b368a0f818dea66eda48be3d"} Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.190050 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b9d7929977db6977ed45ea33dfccc12c6cb2421b368a0f818dea66eda48be3d" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.190116 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202643 4843 generic.go:334] "Generic (PLEG): container finished" podID="5a623a35-5d11-45c3-8874-8993915bb09e" containerID="75fa0c663ae9908d5b20a2bcbcc935c04d1a6446e6cd49da066c6c5a63421cb9" exitCode=0 Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202683 4843 generic.go:334] "Generic (PLEG): container finished" podID="5a623a35-5d11-45c3-8874-8993915bb09e" containerID="4e0d55d20991c5006332a308ead2ac66aac4ab0e0541dc11565c005fd1f2704b" exitCode=2 Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202694 4843 generic.go:334] "Generic (PLEG): container finished" podID="5a623a35-5d11-45c3-8874-8993915bb09e" containerID="8b857a047d7a307b40f75b574496934e8088d289b855493840a40cb1d916287f" exitCode=0 Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202704 4843 generic.go:334] "Generic (PLEG): container finished" podID="5a623a35-5d11-45c3-8874-8993915bb09e" containerID="0c3b43f370bf072d24328cff4e51458308206a6e8e5eabae7a4b17728839746b" exitCode=0 Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202715 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerDied","Data":"75fa0c663ae9908d5b20a2bcbcc935c04d1a6446e6cd49da066c6c5a63421cb9"} Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202757 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerDied","Data":"4e0d55d20991c5006332a308ead2ac66aac4ab0e0541dc11565c005fd1f2704b"} Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202767 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerDied","Data":"8b857a047d7a307b40f75b574496934e8088d289b855493840a40cb1d916287f"} Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.202777 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerDied","Data":"0c3b43f370bf072d24328cff4e51458308206a6e8e5eabae7a4b17728839746b"} Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.285526 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.423865 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-run-httpd\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424192 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-log-httpd\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424219 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-sg-core-conf-yaml\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424242 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424333 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfqk5\" (UniqueName: \"kubernetes.io/projected/5a623a35-5d11-45c3-8874-8993915bb09e-kube-api-access-hfqk5\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424391 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-ceilometer-tls-certs\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424412 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-config-data\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424448 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-combined-ca-bundle\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424464 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-scripts\") pod \"5a623a35-5d11-45c3-8874-8993915bb09e\" (UID: \"5a623a35-5d11-45c3-8874-8993915bb09e\") " Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424681 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424938 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.424966 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a623a35-5d11-45c3-8874-8993915bb09e-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.433502 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-scripts" (OuterVolumeSpecName: "scripts") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.433531 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a623a35-5d11-45c3-8874-8993915bb09e-kube-api-access-hfqk5" (OuterVolumeSpecName: "kube-api-access-hfqk5") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "kube-api-access-hfqk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.447824 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.474515 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.526055 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.526083 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfqk5\" (UniqueName: \"kubernetes.io/projected/5a623a35-5d11-45c3-8874-8993915bb09e-kube-api-access-hfqk5\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.526094 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.526104 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.530984 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-config-data" (OuterVolumeSpecName: "config-data") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.539674 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a623a35-5d11-45c3-8874-8993915bb09e" (UID: "5a623a35-5d11-45c3-8874-8993915bb09e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.627810 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:06 crc kubenswrapper[4843]: I0314 09:53:06.627850 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a623a35-5d11-45c3-8874-8993915bb09e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.224209 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"5a623a35-5d11-45c3-8874-8993915bb09e","Type":"ContainerDied","Data":"0b9e9fdb29deb589536828aff86ba957937638b6e7b15dc7c3cd72abc0d811c8"} Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.224289 4843 scope.go:117] "RemoveContainer" containerID="75fa0c663ae9908d5b20a2bcbcc935c04d1a6446e6cd49da066c6c5a63421cb9" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.224291 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.260658 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.266190 4843 scope.go:117] "RemoveContainer" containerID="4e0d55d20991c5006332a308ead2ac66aac4ab0e0541dc11565c005fd1f2704b" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.269005 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.284773 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:07 crc kubenswrapper[4843]: E0314 09:53:07.285144 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-notification-agent" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285165 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-notification-agent" Mar 14 09:53:07 crc kubenswrapper[4843]: E0314 09:53:07.285185 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="proxy-httpd" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285194 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="proxy-httpd" Mar 14 09:53:07 crc kubenswrapper[4843]: E0314 09:53:07.285214 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" containerName="watcher-db-manage" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285224 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" containerName="watcher-db-manage" Mar 14 09:53:07 crc kubenswrapper[4843]: E0314 09:53:07.285239 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-central-agent" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285246 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-central-agent" Mar 14 09:53:07 crc kubenswrapper[4843]: E0314 09:53:07.285269 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="sg-core" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285294 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="sg-core" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285470 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-notification-agent" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285487 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="sg-core" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285497 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="proxy-httpd" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285513 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" containerName="ceilometer-central-agent" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.285523 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" containerName="watcher-db-manage" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.287235 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.293244 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.293354 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.293609 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.310640 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.312517 4843 scope.go:117] "RemoveContainer" containerID="8b857a047d7a307b40f75b574496934e8088d289b855493840a40cb1d916287f" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339242 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2nvj\" (UniqueName: \"kubernetes.io/projected/b88524af-3d1a-4551-9be0-91c2661bf00b-kube-api-access-p2nvj\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339392 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339492 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-scripts\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339547 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-config-data\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339635 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-log-httpd\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339718 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339792 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.339924 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-run-httpd\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.348238 4843 scope.go:117] "RemoveContainer" containerID="0c3b43f370bf072d24328cff4e51458308206a6e8e5eabae7a4b17728839746b" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.348394 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a623a35-5d11-45c3-8874-8993915bb09e" path="/var/lib/kubelet/pods/5a623a35-5d11-45c3-8874-8993915bb09e/volumes" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.441912 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-run-httpd\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.442390 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-run-httpd\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.442392 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2nvj\" (UniqueName: \"kubernetes.io/projected/b88524af-3d1a-4551-9be0-91c2661bf00b-kube-api-access-p2nvj\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.442464 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.442491 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-scripts\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.442520 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-config-data\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.443048 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-log-httpd\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.443114 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.443166 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.443370 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-log-httpd\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.447656 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.447900 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.447985 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.448531 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-config-data\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.461002 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-scripts\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.462945 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2nvj\" (UniqueName: \"kubernetes.io/projected/b88524af-3d1a-4551-9be0-91c2661bf00b-kube-api-access-p2nvj\") pod \"ceilometer-0\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:07 crc kubenswrapper[4843]: I0314 09:53:07.613922 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:08 crc kubenswrapper[4843]: W0314 09:53:08.139689 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb88524af_3d1a_4551_9be0_91c2661bf00b.slice/crio-d62654f6f04576ab60b2b4eac3aca72ad2a7588501b8544e9ff19a08378e0562 WatchSource:0}: Error finding container d62654f6f04576ab60b2b4eac3aca72ad2a7588501b8544e9ff19a08378e0562: Status 404 returned error can't find the container with id d62654f6f04576ab60b2b4eac3aca72ad2a7588501b8544e9ff19a08378e0562 Mar 14 09:53:08 crc kubenswrapper[4843]: I0314 09:53:08.142912 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:08 crc kubenswrapper[4843]: I0314 09:53:08.235958 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerStarted","Data":"d62654f6f04576ab60b2b4eac3aca72ad2a7588501b8544e9ff19a08378e0562"} Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.245830 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerStarted","Data":"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434"} Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.364338 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-br9mg"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.379714 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-br9mg"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.394928 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.411075 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-purge-29558033-fhmph"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.424028 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-vh296"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.425095 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.430641 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.431130 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" containerName="watcher-applier" containerID="cri-o://e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" gracePeriod=30 Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.439936 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-vh296"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.474131 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74675738-7395-42aa-8156-96e3a9c9c7cb-operator-scripts\") pod \"watchertest-account-delete-vh296\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.474323 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xms9q\" (UniqueName: \"kubernetes.io/projected/74675738-7395-42aa-8156-96e3a9c9c7cb-kube-api-access-xms9q\") pod \"watchertest-account-delete-vh296\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.543306 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.543810 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="6f569cfa-dba2-4b00-8dd9-052336b7014a" containerName="watcher-decision-engine" containerID="cri-o://019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351" gracePeriod=30 Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.556419 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.556706 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-kuttl-api-log" containerID="cri-o://28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82" gracePeriod=30 Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.557143 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-api" containerID="cri-o://f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b" gracePeriod=30 Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.575428 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74675738-7395-42aa-8156-96e3a9c9c7cb-operator-scripts\") pod \"watchertest-account-delete-vh296\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.575530 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xms9q\" (UniqueName: \"kubernetes.io/projected/74675738-7395-42aa-8156-96e3a9c9c7cb-kube-api-access-xms9q\") pod \"watchertest-account-delete-vh296\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.576439 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74675738-7395-42aa-8156-96e3a9c9c7cb-operator-scripts\") pod \"watchertest-account-delete-vh296\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.580557 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.580819 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-kuttl-api-log" containerID="cri-o://924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16" gracePeriod=30 Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.583735 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-api" containerID="cri-o://518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784" gracePeriod=30 Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.613034 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xms9q\" (UniqueName: \"kubernetes.io/projected/74675738-7395-42aa-8156-96e3a9c9c7cb-kube-api-access-xms9q\") pod \"watchertest-account-delete-vh296\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:09 crc kubenswrapper[4843]: I0314 09:53:09.756658 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.110838 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-vh296"] Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.304888 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-vh296" event={"ID":"74675738-7395-42aa-8156-96e3a9c9c7cb","Type":"ContainerStarted","Data":"ed3ebf48bd73d6776daeeaab9d0720e3739f6e05b55211a98d61589153adf2cb"} Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.333516 4843 generic.go:334] "Generic (PLEG): container finished" podID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerID="924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16" exitCode=143 Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.333600 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"447dff52-cadc-414d-9ba3-73d307d1bb35","Type":"ContainerDied","Data":"924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16"} Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.348511 4843 generic.go:334] "Generic (PLEG): container finished" podID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerID="28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82" exitCode=143 Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.348606 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"df4de54e-5485-44e1-80bc-46f18428bb9d","Type":"ContainerDied","Data":"28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82"} Mar 14 09:53:10 crc kubenswrapper[4843]: I0314 09:53:10.354350 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerStarted","Data":"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b"} Mar 14 09:53:11 crc kubenswrapper[4843]: I0314 09:53:11.349110 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="238d64c0-54b7-4541-a7f0-01a95fc2a821" path="/var/lib/kubelet/pods/238d64c0-54b7-4541-a7f0-01a95fc2a821/volumes" Mar 14 09:53:11 crc kubenswrapper[4843]: I0314 09:53:11.349696 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e91ce4b-a255-4a4b-ad9f-0510a8db94b1" path="/var/lib/kubelet/pods/2e91ce4b-a255-4a4b-ad9f-0510a8db94b1/volumes" Mar 14 09:53:11 crc kubenswrapper[4843]: I0314 09:53:11.364948 4843 generic.go:334] "Generic (PLEG): container finished" podID="74675738-7395-42aa-8156-96e3a9c9c7cb" containerID="c7fa90f46e3c117b478bc5938bf0826a9a90208f63cdab683cb88b9a8ee6b649" exitCode=0 Mar 14 09:53:11 crc kubenswrapper[4843]: I0314 09:53:11.365000 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-vh296" event={"ID":"74675738-7395-42aa-8156-96e3a9c9c7cb","Type":"ContainerDied","Data":"c7fa90f46e3c117b478bc5938bf0826a9a90208f63cdab683cb88b9a8ee6b649"} Mar 14 09:53:11 crc kubenswrapper[4843]: I0314 09:53:11.367467 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerStarted","Data":"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e"} Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.054493 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.061114 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.125942 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-combined-ca-bundle\") pod \"447dff52-cadc-414d-9ba3-73d307d1bb35\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126044 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqr4v\" (UniqueName: \"kubernetes.io/projected/df4de54e-5485-44e1-80bc-46f18428bb9d-kube-api-access-vqr4v\") pod \"df4de54e-5485-44e1-80bc-46f18428bb9d\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126068 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df4de54e-5485-44e1-80bc-46f18428bb9d-logs\") pod \"df4de54e-5485-44e1-80bc-46f18428bb9d\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126116 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-config-data\") pod \"447dff52-cadc-414d-9ba3-73d307d1bb35\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126132 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-cert-memcached-mtls\") pod \"df4de54e-5485-44e1-80bc-46f18428bb9d\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126165 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-config-data\") pod \"df4de54e-5485-44e1-80bc-46f18428bb9d\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126190 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qcc7\" (UniqueName: \"kubernetes.io/projected/447dff52-cadc-414d-9ba3-73d307d1bb35-kube-api-access-7qcc7\") pod \"447dff52-cadc-414d-9ba3-73d307d1bb35\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126206 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-custom-prometheus-ca\") pod \"df4de54e-5485-44e1-80bc-46f18428bb9d\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126233 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/447dff52-cadc-414d-9ba3-73d307d1bb35-logs\") pod \"447dff52-cadc-414d-9ba3-73d307d1bb35\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126260 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-combined-ca-bundle\") pod \"df4de54e-5485-44e1-80bc-46f18428bb9d\" (UID: \"df4de54e-5485-44e1-80bc-46f18428bb9d\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126325 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-cert-memcached-mtls\") pod \"447dff52-cadc-414d-9ba3-73d307d1bb35\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.126352 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-custom-prometheus-ca\") pod \"447dff52-cadc-414d-9ba3-73d307d1bb35\" (UID: \"447dff52-cadc-414d-9ba3-73d307d1bb35\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.129972 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df4de54e-5485-44e1-80bc-46f18428bb9d-logs" (OuterVolumeSpecName: "logs") pod "df4de54e-5485-44e1-80bc-46f18428bb9d" (UID: "df4de54e-5485-44e1-80bc-46f18428bb9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.134532 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/447dff52-cadc-414d-9ba3-73d307d1bb35-logs" (OuterVolumeSpecName: "logs") pod "447dff52-cadc-414d-9ba3-73d307d1bb35" (UID: "447dff52-cadc-414d-9ba3-73d307d1bb35"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.138410 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df4de54e-5485-44e1-80bc-46f18428bb9d-kube-api-access-vqr4v" (OuterVolumeSpecName: "kube-api-access-vqr4v") pod "df4de54e-5485-44e1-80bc-46f18428bb9d" (UID: "df4de54e-5485-44e1-80bc-46f18428bb9d"). InnerVolumeSpecName "kube-api-access-vqr4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.155428 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447dff52-cadc-414d-9ba3-73d307d1bb35-kube-api-access-7qcc7" (OuterVolumeSpecName: "kube-api-access-7qcc7") pod "447dff52-cadc-414d-9ba3-73d307d1bb35" (UID: "447dff52-cadc-414d-9ba3-73d307d1bb35"). InnerVolumeSpecName "kube-api-access-7qcc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.180485 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "447dff52-cadc-414d-9ba3-73d307d1bb35" (UID: "447dff52-cadc-414d-9ba3-73d307d1bb35"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.210504 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "447dff52-cadc-414d-9ba3-73d307d1bb35" (UID: "447dff52-cadc-414d-9ba3-73d307d1bb35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.220872 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df4de54e-5485-44e1-80bc-46f18428bb9d" (UID: "df4de54e-5485-44e1-80bc-46f18428bb9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.221100 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "df4de54e-5485-44e1-80bc-46f18428bb9d" (UID: "df4de54e-5485-44e1-80bc-46f18428bb9d"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.222668 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-config-data" (OuterVolumeSpecName: "config-data") pod "df4de54e-5485-44e1-80bc-46f18428bb9d" (UID: "df4de54e-5485-44e1-80bc-46f18428bb9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228426 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228473 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqr4v\" (UniqueName: \"kubernetes.io/projected/df4de54e-5485-44e1-80bc-46f18428bb9d-kube-api-access-vqr4v\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228485 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df4de54e-5485-44e1-80bc-46f18428bb9d-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228493 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228503 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qcc7\" (UniqueName: \"kubernetes.io/projected/447dff52-cadc-414d-9ba3-73d307d1bb35-kube-api-access-7qcc7\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228511 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228519 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/447dff52-cadc-414d-9ba3-73d307d1bb35-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228545 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.228555 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.246588 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-config-data" (OuterVolumeSpecName: "config-data") pod "447dff52-cadc-414d-9ba3-73d307d1bb35" (UID: "447dff52-cadc-414d-9ba3-73d307d1bb35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.253331 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "447dff52-cadc-414d-9ba3-73d307d1bb35" (UID: "447dff52-cadc-414d-9ba3-73d307d1bb35"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.255252 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "df4de54e-5485-44e1-80bc-46f18428bb9d" (UID: "df4de54e-5485-44e1-80bc-46f18428bb9d"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.330143 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.330418 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/df4de54e-5485-44e1-80bc-46f18428bb9d-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.330490 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/447dff52-cadc-414d-9ba3-73d307d1bb35-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.376809 4843 generic.go:334] "Generic (PLEG): container finished" podID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerID="518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784" exitCode=0 Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.376847 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"447dff52-cadc-414d-9ba3-73d307d1bb35","Type":"ContainerDied","Data":"518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784"} Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.376893 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"447dff52-cadc-414d-9ba3-73d307d1bb35","Type":"ContainerDied","Data":"294d930deefc60828f0e690a338446f67d0dbebb3f314be13400c969f4def89b"} Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.376915 4843 scope.go:117] "RemoveContainer" containerID="518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.378070 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.379721 4843 generic.go:334] "Generic (PLEG): container finished" podID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerID="f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b" exitCode=0 Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.380468 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.383362 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"df4de54e-5485-44e1-80bc-46f18428bb9d","Type":"ContainerDied","Data":"f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b"} Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.383401 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"df4de54e-5485-44e1-80bc-46f18428bb9d","Type":"ContainerDied","Data":"5f5888550cab36d7b3fcd903ffc22d261a4276273923ad3f2435e9243f9db3f7"} Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.420882 4843 scope.go:117] "RemoveContainer" containerID="924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.429714 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.445324 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.452377 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.457764 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.471680 4843 scope.go:117] "RemoveContainer" containerID="518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784" Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.472464 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784\": container with ID starting with 518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784 not found: ID does not exist" containerID="518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.472555 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784"} err="failed to get container status \"518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784\": rpc error: code = NotFound desc = could not find container \"518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784\": container with ID starting with 518a33ac74a37146154c0717e85c8f1fa2ca250991af8c27d9e682197d46c784 not found: ID does not exist" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.472628 4843 scope.go:117] "RemoveContainer" containerID="924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16" Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.472981 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16\": container with ID starting with 924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16 not found: ID does not exist" containerID="924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.473065 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16"} err="failed to get container status \"924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16\": rpc error: code = NotFound desc = could not find container \"924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16\": container with ID starting with 924ec109e8b065a55f358e98254b75d7faaa0e01054f58218edcb9bd95c3ae16 not found: ID does not exist" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.473130 4843 scope.go:117] "RemoveContainer" containerID="f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.518967 4843 scope.go:117] "RemoveContainer" containerID="28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.548613 4843 scope.go:117] "RemoveContainer" containerID="f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b" Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.549113 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b\": container with ID starting with f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b not found: ID does not exist" containerID="f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.549150 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b"} err="failed to get container status \"f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b\": rpc error: code = NotFound desc = could not find container \"f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b\": container with ID starting with f7848371a65c2b184335a23d1e1b7582e5bee3946dbee6124a22bd55ed1d1f2b not found: ID does not exist" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.549179 4843 scope.go:117] "RemoveContainer" containerID="28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82" Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.549893 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82\": container with ID starting with 28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82 not found: ID does not exist" containerID="28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.549918 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82"} err="failed to get container status \"28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82\": rpc error: code = NotFound desc = could not find container \"28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82\": container with ID starting with 28c7f27e7614c1c5812a72d87f75a999af4d19cda213ddac48eac68359438a82 not found: ID does not exist" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.761633 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.773795 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.790375 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.803306 4843 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Mar 14 09:53:12 crc kubenswrapper[4843]: E0314 09:53:12.803374 4843 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" containerName="watcher-applier" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.842212 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74675738-7395-42aa-8156-96e3a9c9c7cb-operator-scripts\") pod \"74675738-7395-42aa-8156-96e3a9c9c7cb\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.842834 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74675738-7395-42aa-8156-96e3a9c9c7cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "74675738-7395-42aa-8156-96e3a9c9c7cb" (UID: "74675738-7395-42aa-8156-96e3a9c9c7cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.842930 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xms9q\" (UniqueName: \"kubernetes.io/projected/74675738-7395-42aa-8156-96e3a9c9c7cb-kube-api-access-xms9q\") pod \"74675738-7395-42aa-8156-96e3a9c9c7cb\" (UID: \"74675738-7395-42aa-8156-96e3a9c9c7cb\") " Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.843493 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74675738-7395-42aa-8156-96e3a9c9c7cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.848211 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74675738-7395-42aa-8156-96e3a9c9c7cb-kube-api-access-xms9q" (OuterVolumeSpecName: "kube-api-access-xms9q") pod "74675738-7395-42aa-8156-96e3a9c9c7cb" (UID: "74675738-7395-42aa-8156-96e3a9c9c7cb"). InnerVolumeSpecName "kube-api-access-xms9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.944657 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xms9q\" (UniqueName: \"kubernetes.io/projected/74675738-7395-42aa-8156-96e3a9c9c7cb-kube-api-access-xms9q\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:12 crc kubenswrapper[4843]: I0314 09:53:12.965459 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.348584 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" path="/var/lib/kubelet/pods/447dff52-cadc-414d-9ba3-73d307d1bb35/volumes" Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.349629 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" path="/var/lib/kubelet/pods/df4de54e-5485-44e1-80bc-46f18428bb9d/volumes" Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.389528 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watchertest-account-delete-vh296" Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.389527 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watchertest-account-delete-vh296" event={"ID":"74675738-7395-42aa-8156-96e3a9c9c7cb","Type":"ContainerDied","Data":"ed3ebf48bd73d6776daeeaab9d0720e3739f6e05b55211a98d61589153adf2cb"} Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.389668 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed3ebf48bd73d6776daeeaab9d0720e3739f6e05b55211a98d61589153adf2cb" Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.401284 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerStarted","Data":"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618"} Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.401488 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:13 crc kubenswrapper[4843]: I0314 09:53:13.795669 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.56276348 podStartE2EDuration="6.795652593s" podCreationTimestamp="2026-03-14 09:53:07 +0000 UTC" firstStartedPulling="2026-03-14 09:53:08.142473841 +0000 UTC m=+2555.455084969" lastFinishedPulling="2026-03-14 09:53:12.375362954 +0000 UTC m=+2559.687974082" observedRunningTime="2026-03-14 09:53:13.426951334 +0000 UTC m=+2560.739562462" watchObservedRunningTime="2026-03-14 09:53:13.795652593 +0000 UTC m=+2561.108263721" Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.408983 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-central-agent" containerID="cri-o://783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" gracePeriod=30 Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.409480 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="proxy-httpd" containerID="cri-o://61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" gracePeriod=30 Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.409535 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="sg-core" containerID="cri-o://b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" gracePeriod=30 Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.409567 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-notification-agent" containerID="cri-o://c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" gracePeriod=30 Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.443846 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-qwqwz"] Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.451947 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-qwqwz"] Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.458739 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-vh296"] Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.464424 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-gd7kp"] Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.469663 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watchertest-account-delete-vh296"] Mar 14 09:53:14 crc kubenswrapper[4843]: I0314 09:53:14.474833 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-test-account-create-update-gd7kp"] Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.096453 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.171846 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.181078 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-config-data\") pod \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.181169 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-combined-ca-bundle\") pod \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.181200 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-cert-memcached-mtls\") pod \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.181376 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0f0536-eea6-4e03-86a4-f3a077128d0e-logs\") pod \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.181405 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg4q6\" (UniqueName: \"kubernetes.io/projected/6a0f0536-eea6-4e03-86a4-f3a077128d0e-kube-api-access-sg4q6\") pod \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\" (UID: \"6a0f0536-eea6-4e03-86a4-f3a077128d0e\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.181899 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a0f0536-eea6-4e03-86a4-f3a077128d0e-logs" (OuterVolumeSpecName: "logs") pod "6a0f0536-eea6-4e03-86a4-f3a077128d0e" (UID: "6a0f0536-eea6-4e03-86a4-f3a077128d0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.193077 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0f0536-eea6-4e03-86a4-f3a077128d0e-kube-api-access-sg4q6" (OuterVolumeSpecName: "kube-api-access-sg4q6") pod "6a0f0536-eea6-4e03-86a4-f3a077128d0e" (UID: "6a0f0536-eea6-4e03-86a4-f3a077128d0e"). InnerVolumeSpecName "kube-api-access-sg4q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.231509 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a0f0536-eea6-4e03-86a4-f3a077128d0e" (UID: "6a0f0536-eea6-4e03-86a4-f3a077128d0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.237924 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-config-data" (OuterVolumeSpecName: "config-data") pod "6a0f0536-eea6-4e03-86a4-f3a077128d0e" (UID: "6a0f0536-eea6-4e03-86a4-f3a077128d0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.283222 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-combined-ca-bundle\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.286927 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-ceilometer-tls-certs\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.286988 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2nvj\" (UniqueName: \"kubernetes.io/projected/b88524af-3d1a-4551-9be0-91c2661bf00b-kube-api-access-p2nvj\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287224 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-log-httpd\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287309 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-sg-core-conf-yaml\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287344 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-scripts\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287373 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-run-httpd\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287418 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-config-data\") pod \"b88524af-3d1a-4551-9be0-91c2661bf00b\" (UID: \"b88524af-3d1a-4551-9be0-91c2661bf00b\") " Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287701 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287912 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287935 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287950 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0f0536-eea6-4e03-86a4-f3a077128d0e-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287961 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg4q6\" (UniqueName: \"kubernetes.io/projected/6a0f0536-eea6-4e03-86a4-f3a077128d0e-kube-api-access-sg4q6\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.287973 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.288398 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.292461 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-scripts" (OuterVolumeSpecName: "scripts") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.297220 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b88524af-3d1a-4551-9be0-91c2661bf00b-kube-api-access-p2nvj" (OuterVolumeSpecName: "kube-api-access-p2nvj") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "kube-api-access-p2nvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.317527 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "6a0f0536-eea6-4e03-86a4-f3a077128d0e" (UID: "6a0f0536-eea6-4e03-86a4-f3a077128d0e"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.330163 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.348987 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.361091 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74675738-7395-42aa-8156-96e3a9c9c7cb" path="/var/lib/kubelet/pods/74675738-7395-42aa-8156-96e3a9c9c7cb/volumes" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.361644 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d3682d7-b738-4ca1-9a9a-10b4d5fde44a" path="/var/lib/kubelet/pods/8d3682d7-b738-4ca1-9a9a-10b4d5fde44a/volumes" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.362479 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ab63c1-c44d-4957-bdc7-5b3fd7e92795" path="/var/lib/kubelet/pods/c8ab63c1-c44d-4957-bdc7-5b3fd7e92795/volumes" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.375991 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-config-data" (OuterVolumeSpecName: "config-data") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.378198 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b88524af-3d1a-4551-9be0-91c2661bf00b" (UID: "b88524af-3d1a-4551-9be0-91c2661bf00b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.389552 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b88524af-3d1a-4551-9be0-91c2661bf00b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.389871 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.389966 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.390028 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.390091 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.390151 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6a0f0536-eea6-4e03-86a4-f3a077128d0e-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.390207 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b88524af-3d1a-4551-9be0-91c2661bf00b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.390298 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2nvj\" (UniqueName: \"kubernetes.io/projected/b88524af-3d1a-4551-9be0-91c2661bf00b-kube-api-access-p2nvj\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.419284 4843 generic.go:334] "Generic (PLEG): container finished" podID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" exitCode=0 Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.419373 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.420432 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6a0f0536-eea6-4e03-86a4-f3a077128d0e","Type":"ContainerDied","Data":"e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.420456 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"6a0f0536-eea6-4e03-86a4-f3a077128d0e","Type":"ContainerDied","Data":"14b34d60805f1b2b9cdb48622e6b79674f24f9aa6a091aa2a1fc0e5c4a54a86c"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.420471 4843 scope.go:117] "RemoveContainer" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424845 4843 generic.go:334] "Generic (PLEG): container finished" podID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" exitCode=0 Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424887 4843 generic.go:334] "Generic (PLEG): container finished" podID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" exitCode=2 Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424895 4843 generic.go:334] "Generic (PLEG): container finished" podID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" exitCode=0 Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424902 4843 generic.go:334] "Generic (PLEG): container finished" podID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" exitCode=0 Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424921 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerDied","Data":"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424945 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerDied","Data":"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424955 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerDied","Data":"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424964 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerDied","Data":"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.424974 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"b88524af-3d1a-4551-9be0-91c2661bf00b","Type":"ContainerDied","Data":"d62654f6f04576ab60b2b4eac3aca72ad2a7588501b8544e9ff19a08378e0562"} Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.425053 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.446833 4843 scope.go:117] "RemoveContainer" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.447736 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0\": container with ID starting with e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0 not found: ID does not exist" containerID="e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.447779 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0"} err="failed to get container status \"e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0\": rpc error: code = NotFound desc = could not find container \"e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0\": container with ID starting with e4ca84c7627ca1ffb37c40c054dd442a84dfae5557772ed9ac753ed320d8b7b0 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.447803 4843 scope.go:117] "RemoveContainer" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.451108 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.456401 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.472247 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.474558 4843 scope.go:117] "RemoveContainer" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.481956 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.506878 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507343 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-api" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507367 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-api" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507388 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="sg-core" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507397 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="sg-core" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507410 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-kuttl-api-log" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507419 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-kuttl-api-log" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507445 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-api" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507453 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-api" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507465 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74675738-7395-42aa-8156-96e3a9c9c7cb" containerName="mariadb-account-delete" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507472 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="74675738-7395-42aa-8156-96e3a9c9c7cb" containerName="mariadb-account-delete" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507491 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-kuttl-api-log" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507500 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-kuttl-api-log" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507513 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" containerName="watcher-applier" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507521 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" containerName="watcher-applier" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507531 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="proxy-httpd" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507538 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="proxy-httpd" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507549 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-notification-agent" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507557 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-notification-agent" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.507569 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-central-agent" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507576 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-central-agent" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507764 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="proxy-httpd" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507789 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-api" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507798 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-api" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507808 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-notification-agent" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507815 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="447dff52-cadc-414d-9ba3-73d307d1bb35" containerName="watcher-kuttl-api-log" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507826 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="df4de54e-5485-44e1-80bc-46f18428bb9d" containerName="watcher-kuttl-api-log" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507840 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="sg-core" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507851 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="74675738-7395-42aa-8156-96e3a9c9c7cb" containerName="mariadb-account-delete" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507865 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" containerName="ceilometer-central-agent" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.507874 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" containerName="watcher-applier" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.510801 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.512053 4843 scope.go:117] "RemoveContainer" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.515637 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.515848 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.515995 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.523000 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.595517 4843 scope.go:117] "RemoveContainer" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596446 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596545 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-scripts\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596596 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596642 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-config-data\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596665 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-log-httpd\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596767 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596802 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q8w2\" (UniqueName: \"kubernetes.io/projected/42c05d98-1fe3-4a70-ac34-8f3601433809-kube-api-access-6q8w2\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.596824 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-run-httpd\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.620329 4843 scope.go:117] "RemoveContainer" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.620839 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": container with ID starting with 61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618 not found: ID does not exist" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.620880 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618"} err="failed to get container status \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": rpc error: code = NotFound desc = could not find container \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": container with ID starting with 61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.620919 4843 scope.go:117] "RemoveContainer" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.621416 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": container with ID starting with b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e not found: ID does not exist" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.621457 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e"} err="failed to get container status \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": rpc error: code = NotFound desc = could not find container \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": container with ID starting with b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.621470 4843 scope.go:117] "RemoveContainer" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.621658 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": container with ID starting with c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b not found: ID does not exist" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.621694 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b"} err="failed to get container status \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": rpc error: code = NotFound desc = could not find container \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": container with ID starting with c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.621709 4843 scope.go:117] "RemoveContainer" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" Mar 14 09:53:15 crc kubenswrapper[4843]: E0314 09:53:15.621947 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": container with ID starting with 783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434 not found: ID does not exist" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.621969 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434"} err="failed to get container status \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": rpc error: code = NotFound desc = could not find container \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": container with ID starting with 783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.621980 4843 scope.go:117] "RemoveContainer" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.622214 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618"} err="failed to get container status \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": rpc error: code = NotFound desc = could not find container \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": container with ID starting with 61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.622256 4843 scope.go:117] "RemoveContainer" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.622745 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e"} err="failed to get container status \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": rpc error: code = NotFound desc = could not find container \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": container with ID starting with b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.622786 4843 scope.go:117] "RemoveContainer" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623060 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b"} err="failed to get container status \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": rpc error: code = NotFound desc = could not find container \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": container with ID starting with c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623092 4843 scope.go:117] "RemoveContainer" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623338 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434"} err="failed to get container status \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": rpc error: code = NotFound desc = could not find container \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": container with ID starting with 783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623362 4843 scope.go:117] "RemoveContainer" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623569 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618"} err="failed to get container status \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": rpc error: code = NotFound desc = could not find container \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": container with ID starting with 61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623603 4843 scope.go:117] "RemoveContainer" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623894 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e"} err="failed to get container status \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": rpc error: code = NotFound desc = could not find container \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": container with ID starting with b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.623925 4843 scope.go:117] "RemoveContainer" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624167 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b"} err="failed to get container status \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": rpc error: code = NotFound desc = could not find container \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": container with ID starting with c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624189 4843 scope.go:117] "RemoveContainer" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624374 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434"} err="failed to get container status \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": rpc error: code = NotFound desc = could not find container \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": container with ID starting with 783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624388 4843 scope.go:117] "RemoveContainer" containerID="61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624602 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618"} err="failed to get container status \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": rpc error: code = NotFound desc = could not find container \"61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618\": container with ID starting with 61d2564840b61b47e88855faac06129b2f0a4202b014f1015ed92043fba2c618 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624648 4843 scope.go:117] "RemoveContainer" containerID="b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.624984 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e"} err="failed to get container status \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": rpc error: code = NotFound desc = could not find container \"b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e\": container with ID starting with b26e315bcaf5c3b15dee13061f9973058d4a48a0c59a6fa5e80f248a3106025e not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.625006 4843 scope.go:117] "RemoveContainer" containerID="c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.625243 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b"} err="failed to get container status \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": rpc error: code = NotFound desc = could not find container \"c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b\": container with ID starting with c4fff99d9d10e55598fb184417e03b3b86f581d8b0ffaa7d704cc34297ea795b not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.625265 4843 scope.go:117] "RemoveContainer" containerID="783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.625599 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434"} err="failed to get container status \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": rpc error: code = NotFound desc = could not find container \"783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434\": container with ID starting with 783a70ee4e0cc30c0d0db10534368c2de4069eb1e52c7279a82d6bcfe5d88434 not found: ID does not exist" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698073 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698126 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q8w2\" (UniqueName: \"kubernetes.io/projected/42c05d98-1fe3-4a70-ac34-8f3601433809-kube-api-access-6q8w2\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698149 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-run-httpd\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698186 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698236 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-scripts\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698319 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698345 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-config-data\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.698364 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-log-httpd\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.699149 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-log-httpd\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.699329 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-run-httpd\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.702664 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.703661 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-scripts\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.704895 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.705554 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.706366 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-config-data\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.715792 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q8w2\" (UniqueName: \"kubernetes.io/projected/42c05d98-1fe3-4a70-ac34-8f3601433809-kube-api-access-6q8w2\") pod \"ceilometer-0\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:15 crc kubenswrapper[4843]: I0314 09:53:15.893006 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.013835 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.108312 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f569cfa-dba2-4b00-8dd9-052336b7014a-logs\") pod \"6f569cfa-dba2-4b00-8dd9-052336b7014a\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.108482 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-cert-memcached-mtls\") pod \"6f569cfa-dba2-4b00-8dd9-052336b7014a\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.108508 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-config-data\") pod \"6f569cfa-dba2-4b00-8dd9-052336b7014a\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.108548 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-combined-ca-bundle\") pod \"6f569cfa-dba2-4b00-8dd9-052336b7014a\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.108610 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-custom-prometheus-ca\") pod \"6f569cfa-dba2-4b00-8dd9-052336b7014a\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.108667 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp2j2\" (UniqueName: \"kubernetes.io/projected/6f569cfa-dba2-4b00-8dd9-052336b7014a-kube-api-access-dp2j2\") pod \"6f569cfa-dba2-4b00-8dd9-052336b7014a\" (UID: \"6f569cfa-dba2-4b00-8dd9-052336b7014a\") " Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.109176 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f569cfa-dba2-4b00-8dd9-052336b7014a-logs" (OuterVolumeSpecName: "logs") pod "6f569cfa-dba2-4b00-8dd9-052336b7014a" (UID: "6f569cfa-dba2-4b00-8dd9-052336b7014a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.113591 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f569cfa-dba2-4b00-8dd9-052336b7014a-kube-api-access-dp2j2" (OuterVolumeSpecName: "kube-api-access-dp2j2") pod "6f569cfa-dba2-4b00-8dd9-052336b7014a" (UID: "6f569cfa-dba2-4b00-8dd9-052336b7014a"). InnerVolumeSpecName "kube-api-access-dp2j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.133946 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f569cfa-dba2-4b00-8dd9-052336b7014a" (UID: "6f569cfa-dba2-4b00-8dd9-052336b7014a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.136503 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "6f569cfa-dba2-4b00-8dd9-052336b7014a" (UID: "6f569cfa-dba2-4b00-8dd9-052336b7014a"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.165947 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-config-data" (OuterVolumeSpecName: "config-data") pod "6f569cfa-dba2-4b00-8dd9-052336b7014a" (UID: "6f569cfa-dba2-4b00-8dd9-052336b7014a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.192455 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "6f569cfa-dba2-4b00-8dd9-052336b7014a" (UID: "6f569cfa-dba2-4b00-8dd9-052336b7014a"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.211031 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp2j2\" (UniqueName: \"kubernetes.io/projected/6f569cfa-dba2-4b00-8dd9-052336b7014a-kube-api-access-dp2j2\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.211078 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f569cfa-dba2-4b00-8dd9-052336b7014a-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.211093 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.211104 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.211115 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.211127 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/6f569cfa-dba2-4b00-8dd9-052336b7014a-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.350964 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.437080 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerStarted","Data":"bc9dc11f44cb1ab237cd6dd5450c1c5da73437461af281f9e7ea23b73d9d4236"} Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.438809 4843 generic.go:334] "Generic (PLEG): container finished" podID="6f569cfa-dba2-4b00-8dd9-052336b7014a" containerID="019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351" exitCode=0 Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.438863 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"6f569cfa-dba2-4b00-8dd9-052336b7014a","Type":"ContainerDied","Data":"019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351"} Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.438881 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"6f569cfa-dba2-4b00-8dd9-052336b7014a","Type":"ContainerDied","Data":"b2c6df07f138bf153032f1534b356e5512ec4b661368f2c5a888bc78550efb6e"} Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.438900 4843 scope.go:117] "RemoveContainer" containerID="019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.438901 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.489454 4843 scope.go:117] "RemoveContainer" containerID="019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351" Mar 14 09:53:16 crc kubenswrapper[4843]: E0314 09:53:16.491049 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351\": container with ID starting with 019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351 not found: ID does not exist" containerID="019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.491127 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351"} err="failed to get container status \"019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351\": rpc error: code = NotFound desc = could not find container \"019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351\": container with ID starting with 019cb68a545845d143cc7a27a4f092d96cb51cdd6e2ac981f4b37aac725e1351 not found: ID does not exist" Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.491785 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:53:16 crc kubenswrapper[4843]: I0314 09:53:16.503534 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:53:17 crc kubenswrapper[4843]: I0314 09:53:17.354210 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a0f0536-eea6-4e03-86a4-f3a077128d0e" path="/var/lib/kubelet/pods/6a0f0536-eea6-4e03-86a4-f3a077128d0e/volumes" Mar 14 09:53:17 crc kubenswrapper[4843]: I0314 09:53:17.355293 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f569cfa-dba2-4b00-8dd9-052336b7014a" path="/var/lib/kubelet/pods/6f569cfa-dba2-4b00-8dd9-052336b7014a/volumes" Mar 14 09:53:17 crc kubenswrapper[4843]: I0314 09:53:17.355947 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b88524af-3d1a-4551-9be0-91c2661bf00b" path="/var/lib/kubelet/pods/b88524af-3d1a-4551-9be0-91c2661bf00b/volumes" Mar 14 09:53:17 crc kubenswrapper[4843]: I0314 09:53:17.462194 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerStarted","Data":"307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437"} Mar 14 09:53:18 crc kubenswrapper[4843]: I0314 09:53:18.475481 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerStarted","Data":"8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76"} Mar 14 09:53:18 crc kubenswrapper[4843]: I0314 09:53:18.475915 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerStarted","Data":"4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1"} Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.156230 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-q467f"] Mar 14 09:53:19 crc kubenswrapper[4843]: E0314 09:53:19.156547 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f569cfa-dba2-4b00-8dd9-052336b7014a" containerName="watcher-decision-engine" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.156559 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f569cfa-dba2-4b00-8dd9-052336b7014a" containerName="watcher-decision-engine" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.156695 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f569cfa-dba2-4b00-8dd9-052336b7014a" containerName="watcher-decision-engine" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.157187 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.170665 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-q467f"] Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.240391 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-1522-account-create-update-brbfc"] Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.241406 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.246436 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-1522-account-create-update-brbfc"] Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.248042 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.260815 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d271cc6-c547-4e3d-83a3-248d29ace879-operator-scripts\") pod \"watcher-db-create-q467f\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.261051 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kql7m\" (UniqueName: \"kubernetes.io/projected/0d271cc6-c547-4e3d-83a3-248d29ace879-kube-api-access-kql7m\") pod \"watcher-db-create-q467f\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.363075 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26d49\" (UniqueName: \"kubernetes.io/projected/ea5456e0-feac-43fe-b412-46cfe1cff4de-kube-api-access-26d49\") pod \"watcher-1522-account-create-update-brbfc\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.363161 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d271cc6-c547-4e3d-83a3-248d29ace879-operator-scripts\") pod \"watcher-db-create-q467f\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.363218 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kql7m\" (UniqueName: \"kubernetes.io/projected/0d271cc6-c547-4e3d-83a3-248d29ace879-kube-api-access-kql7m\") pod \"watcher-db-create-q467f\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.363311 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea5456e0-feac-43fe-b412-46cfe1cff4de-operator-scripts\") pod \"watcher-1522-account-create-update-brbfc\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.363952 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d271cc6-c547-4e3d-83a3-248d29ace879-operator-scripts\") pod \"watcher-db-create-q467f\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.380688 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kql7m\" (UniqueName: \"kubernetes.io/projected/0d271cc6-c547-4e3d-83a3-248d29ace879-kube-api-access-kql7m\") pod \"watcher-db-create-q467f\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.465088 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26d49\" (UniqueName: \"kubernetes.io/projected/ea5456e0-feac-43fe-b412-46cfe1cff4de-kube-api-access-26d49\") pod \"watcher-1522-account-create-update-brbfc\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.465330 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea5456e0-feac-43fe-b412-46cfe1cff4de-operator-scripts\") pod \"watcher-1522-account-create-update-brbfc\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.466810 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea5456e0-feac-43fe-b412-46cfe1cff4de-operator-scripts\") pod \"watcher-1522-account-create-update-brbfc\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.472480 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.502032 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26d49\" (UniqueName: \"kubernetes.io/projected/ea5456e0-feac-43fe-b412-46cfe1cff4de-kube-api-access-26d49\") pod \"watcher-1522-account-create-update-brbfc\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:19 crc kubenswrapper[4843]: I0314 09:53:19.557811 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.019403 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-q467f"] Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.223659 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-1522-account-create-update-brbfc"] Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.514476 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerStarted","Data":"f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae"} Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.514802 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.516913 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" event={"ID":"ea5456e0-feac-43fe-b412-46cfe1cff4de","Type":"ContainerStarted","Data":"22ee577c0712d7b442ae0692d0bed7d005fd215b2038b7fb1ad7354aaeba1bcd"} Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.516981 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" event={"ID":"ea5456e0-feac-43fe-b412-46cfe1cff4de","Type":"ContainerStarted","Data":"16adb6954d176d00f37ce6cc7baff8032f633e096ba7b138dad6ca389c84824b"} Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.518788 4843 generic.go:334] "Generic (PLEG): container finished" podID="0d271cc6-c547-4e3d-83a3-248d29ace879" containerID="071ff12ce7fb8b9e252e7d5604e64782d0a54b63f6edec35ce091e547c16d716" exitCode=0 Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.518815 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-q467f" event={"ID":"0d271cc6-c547-4e3d-83a3-248d29ace879","Type":"ContainerDied","Data":"071ff12ce7fb8b9e252e7d5604e64782d0a54b63f6edec35ce091e547c16d716"} Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.518833 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-q467f" event={"ID":"0d271cc6-c547-4e3d-83a3-248d29ace879","Type":"ContainerStarted","Data":"a5f95f52119da660954d8b7f9d105c638e663632cc89b5357ddf79b754553b42"} Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.540287 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.099502545 podStartE2EDuration="5.540263072s" podCreationTimestamp="2026-03-14 09:53:15 +0000 UTC" firstStartedPulling="2026-03-14 09:53:16.35686768 +0000 UTC m=+2563.669478808" lastFinishedPulling="2026-03-14 09:53:19.797628207 +0000 UTC m=+2567.110239335" observedRunningTime="2026-03-14 09:53:20.539571595 +0000 UTC m=+2567.852182723" watchObservedRunningTime="2026-03-14 09:53:20.540263072 +0000 UTC m=+2567.852874200" Mar 14 09:53:20 crc kubenswrapper[4843]: I0314 09:53:20.553997 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" podStartSLOduration=1.553976258 podStartE2EDuration="1.553976258s" podCreationTimestamp="2026-03-14 09:53:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:20.551518407 +0000 UTC m=+2567.864129535" watchObservedRunningTime="2026-03-14 09:53:20.553976258 +0000 UTC m=+2567.866587386" Mar 14 09:53:21 crc kubenswrapper[4843]: I0314 09:53:21.530721 4843 generic.go:334] "Generic (PLEG): container finished" podID="ea5456e0-feac-43fe-b412-46cfe1cff4de" containerID="22ee577c0712d7b442ae0692d0bed7d005fd215b2038b7fb1ad7354aaeba1bcd" exitCode=0 Mar 14 09:53:21 crc kubenswrapper[4843]: I0314 09:53:21.530792 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" event={"ID":"ea5456e0-feac-43fe-b412-46cfe1cff4de","Type":"ContainerDied","Data":"22ee577c0712d7b442ae0692d0bed7d005fd215b2038b7fb1ad7354aaeba1bcd"} Mar 14 09:53:21 crc kubenswrapper[4843]: I0314 09:53:21.942378 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.017697 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kql7m\" (UniqueName: \"kubernetes.io/projected/0d271cc6-c547-4e3d-83a3-248d29ace879-kube-api-access-kql7m\") pod \"0d271cc6-c547-4e3d-83a3-248d29ace879\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.017922 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d271cc6-c547-4e3d-83a3-248d29ace879-operator-scripts\") pod \"0d271cc6-c547-4e3d-83a3-248d29ace879\" (UID: \"0d271cc6-c547-4e3d-83a3-248d29ace879\") " Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.018355 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d271cc6-c547-4e3d-83a3-248d29ace879-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d271cc6-c547-4e3d-83a3-248d29ace879" (UID: "0d271cc6-c547-4e3d-83a3-248d29ace879"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.033765 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d271cc6-c547-4e3d-83a3-248d29ace879-kube-api-access-kql7m" (OuterVolumeSpecName: "kube-api-access-kql7m") pod "0d271cc6-c547-4e3d-83a3-248d29ace879" (UID: "0d271cc6-c547-4e3d-83a3-248d29ace879"). InnerVolumeSpecName "kube-api-access-kql7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.119659 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kql7m\" (UniqueName: \"kubernetes.io/projected/0d271cc6-c547-4e3d-83a3-248d29ace879-kube-api-access-kql7m\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.119712 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d271cc6-c547-4e3d-83a3-248d29ace879-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.539902 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-q467f" Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.539878 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-q467f" event={"ID":"0d271cc6-c547-4e3d-83a3-248d29ace879","Type":"ContainerDied","Data":"a5f95f52119da660954d8b7f9d105c638e663632cc89b5357ddf79b754553b42"} Mar 14 09:53:22 crc kubenswrapper[4843]: I0314 09:53:22.540233 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5f95f52119da660954d8b7f9d105c638e663632cc89b5357ddf79b754553b42" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.002145 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.134901 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26d49\" (UniqueName: \"kubernetes.io/projected/ea5456e0-feac-43fe-b412-46cfe1cff4de-kube-api-access-26d49\") pod \"ea5456e0-feac-43fe-b412-46cfe1cff4de\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.135205 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea5456e0-feac-43fe-b412-46cfe1cff4de-operator-scripts\") pod \"ea5456e0-feac-43fe-b412-46cfe1cff4de\" (UID: \"ea5456e0-feac-43fe-b412-46cfe1cff4de\") " Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.135688 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea5456e0-feac-43fe-b412-46cfe1cff4de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea5456e0-feac-43fe-b412-46cfe1cff4de" (UID: "ea5456e0-feac-43fe-b412-46cfe1cff4de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.135883 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea5456e0-feac-43fe-b412-46cfe1cff4de-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.144844 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea5456e0-feac-43fe-b412-46cfe1cff4de-kube-api-access-26d49" (OuterVolumeSpecName: "kube-api-access-26d49") pod "ea5456e0-feac-43fe-b412-46cfe1cff4de" (UID: "ea5456e0-feac-43fe-b412-46cfe1cff4de"). InnerVolumeSpecName "kube-api-access-26d49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.237323 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26d49\" (UniqueName: \"kubernetes.io/projected/ea5456e0-feac-43fe-b412-46cfe1cff4de-kube-api-access-26d49\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.556581 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" event={"ID":"ea5456e0-feac-43fe-b412-46cfe1cff4de","Type":"ContainerDied","Data":"16adb6954d176d00f37ce6cc7baff8032f633e096ba7b138dad6ca389c84824b"} Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.556629 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16adb6954d176d00f37ce6cc7baff8032f633e096ba7b138dad6ca389c84824b" Mar 14 09:53:23 crc kubenswrapper[4843]: I0314 09:53:23.556691 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-1522-account-create-update-brbfc" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.474382 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-8skbb"] Mar 14 09:53:24 crc kubenswrapper[4843]: E0314 09:53:24.475003 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d271cc6-c547-4e3d-83a3-248d29ace879" containerName="mariadb-database-create" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.475016 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d271cc6-c547-4e3d-83a3-248d29ace879" containerName="mariadb-database-create" Mar 14 09:53:24 crc kubenswrapper[4843]: E0314 09:53:24.475045 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5456e0-feac-43fe-b412-46cfe1cff4de" containerName="mariadb-account-create-update" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.475053 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5456e0-feac-43fe-b412-46cfe1cff4de" containerName="mariadb-account-create-update" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.475224 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea5456e0-feac-43fe-b412-46cfe1cff4de" containerName="mariadb-account-create-update" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.475247 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d271cc6-c547-4e3d-83a3-248d29ace879" containerName="mariadb-database-create" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.475906 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.478168 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.479577 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-dzfjc" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.487105 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-8skbb"] Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.564409 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-576dl\" (UniqueName: \"kubernetes.io/projected/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-kube-api-access-576dl\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.565495 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-config-data\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.565750 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-db-sync-config-data\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.565838 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.669396 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-576dl\" (UniqueName: \"kubernetes.io/projected/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-kube-api-access-576dl\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.669448 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-config-data\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.669466 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-db-sync-config-data\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.669497 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.675796 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-config-data\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.676469 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.676773 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-db-sync-config-data\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.686522 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-576dl\" (UniqueName: \"kubernetes.io/projected/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-kube-api-access-576dl\") pod \"watcher-kuttl-db-sync-8skbb\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:24 crc kubenswrapper[4843]: I0314 09:53:24.805749 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:25 crc kubenswrapper[4843]: I0314 09:53:25.295944 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-8skbb"] Mar 14 09:53:25 crc kubenswrapper[4843]: I0314 09:53:25.574513 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" event={"ID":"bc2408d8-efbd-44b0-a7cd-98a5891f7fab","Type":"ContainerStarted","Data":"bb88982aee7a576286d16593e2eda1f05ca76fe570bdc0aa6315c4d00af5508d"} Mar 14 09:53:25 crc kubenswrapper[4843]: I0314 09:53:25.574567 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" event={"ID":"bc2408d8-efbd-44b0-a7cd-98a5891f7fab","Type":"ContainerStarted","Data":"280a028d993e7815a34298b27c0cb1a6049b94650973758b78d4b1fcdc537110"} Mar 14 09:53:25 crc kubenswrapper[4843]: I0314 09:53:25.589408 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" podStartSLOduration=1.589390693 podStartE2EDuration="1.589390693s" podCreationTimestamp="2026-03-14 09:53:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:25.587862295 +0000 UTC m=+2572.900473413" watchObservedRunningTime="2026-03-14 09:53:25.589390693 +0000 UTC m=+2572.902001821" Mar 14 09:53:28 crc kubenswrapper[4843]: I0314 09:53:28.605071 4843 generic.go:334] "Generic (PLEG): container finished" podID="bc2408d8-efbd-44b0-a7cd-98a5891f7fab" containerID="bb88982aee7a576286d16593e2eda1f05ca76fe570bdc0aa6315c4d00af5508d" exitCode=0 Mar 14 09:53:28 crc kubenswrapper[4843]: I0314 09:53:28.605159 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" event={"ID":"bc2408d8-efbd-44b0-a7cd-98a5891f7fab","Type":"ContainerDied","Data":"bb88982aee7a576286d16593e2eda1f05ca76fe570bdc0aa6315c4d00af5508d"} Mar 14 09:53:29 crc kubenswrapper[4843]: I0314 09:53:29.987888 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.054192 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-db-sync-config-data\") pod \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.054334 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-config-data\") pod \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.054371 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-combined-ca-bundle\") pod \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.054419 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-576dl\" (UniqueName: \"kubernetes.io/projected/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-kube-api-access-576dl\") pod \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\" (UID: \"bc2408d8-efbd-44b0-a7cd-98a5891f7fab\") " Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.059764 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-kube-api-access-576dl" (OuterVolumeSpecName: "kube-api-access-576dl") pod "bc2408d8-efbd-44b0-a7cd-98a5891f7fab" (UID: "bc2408d8-efbd-44b0-a7cd-98a5891f7fab"). InnerVolumeSpecName "kube-api-access-576dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.060397 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bc2408d8-efbd-44b0-a7cd-98a5891f7fab" (UID: "bc2408d8-efbd-44b0-a7cd-98a5891f7fab"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.078654 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc2408d8-efbd-44b0-a7cd-98a5891f7fab" (UID: "bc2408d8-efbd-44b0-a7cd-98a5891f7fab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.100495 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-config-data" (OuterVolumeSpecName: "config-data") pod "bc2408d8-efbd-44b0-a7cd-98a5891f7fab" (UID: "bc2408d8-efbd-44b0-a7cd-98a5891f7fab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.155883 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.155918 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.155929 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-576dl\" (UniqueName: \"kubernetes.io/projected/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-kube-api-access-576dl\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.155938 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc2408d8-efbd-44b0-a7cd-98a5891f7fab-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.628419 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" event={"ID":"bc2408d8-efbd-44b0-a7cd-98a5891f7fab","Type":"ContainerDied","Data":"280a028d993e7815a34298b27c0cb1a6049b94650973758b78d4b1fcdc537110"} Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.628479 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="280a028d993e7815a34298b27c0cb1a6049b94650973758b78d4b1fcdc537110" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.628574 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-8skbb" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.936082 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:53:30 crc kubenswrapper[4843]: E0314 09:53:30.943796 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2408d8-efbd-44b0-a7cd-98a5891f7fab" containerName="watcher-kuttl-db-sync" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.943830 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2408d8-efbd-44b0-a7cd-98a5891f7fab" containerName="watcher-kuttl-db-sync" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.944007 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2408d8-efbd-44b0-a7cd-98a5891f7fab" containerName="watcher-kuttl-db-sync" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.944994 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.945762 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.946461 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.948124 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-dzfjc" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.948847 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.957014 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.962596 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:53:30 crc kubenswrapper[4843]: I0314 09:53:30.968002 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.042452 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.043814 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.045918 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.051458 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069283 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vctnx\" (UniqueName: \"kubernetes.io/projected/53784836-318f-4a01-899d-f3cd6cbd72d0-kube-api-access-vctnx\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069344 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069379 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53784836-318f-4a01-899d-f3cd6cbd72d0-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069438 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069562 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069683 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069704 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069745 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069760 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069854 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftxbv\" (UniqueName: \"kubernetes.io/projected/d4524e41-3177-4936-b995-6b0231394c16-kube-api-access-ftxbv\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.069878 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4524e41-3177-4936-b995-6b0231394c16-logs\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.170874 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c847b0bd-b26d-4d23-a465-15d57f1a968e-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.170931 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h8lz\" (UniqueName: \"kubernetes.io/projected/c847b0bd-b26d-4d23-a465-15d57f1a968e-kube-api-access-2h8lz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.170962 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53784836-318f-4a01-899d-f3cd6cbd72d0-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.170986 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171048 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171097 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171148 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171170 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171199 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171227 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171249 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171301 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171337 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftxbv\" (UniqueName: \"kubernetes.io/projected/d4524e41-3177-4936-b995-6b0231394c16-kube-api-access-ftxbv\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171362 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4524e41-3177-4936-b995-6b0231394c16-logs\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171401 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vctnx\" (UniqueName: \"kubernetes.io/projected/53784836-318f-4a01-899d-f3cd6cbd72d0-kube-api-access-vctnx\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171438 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.171466 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.172333 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53784836-318f-4a01-899d-f3cd6cbd72d0-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.173048 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4524e41-3177-4936-b995-6b0231394c16-logs\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.176332 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.176633 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.177011 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.177168 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.179114 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.182194 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.183712 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.191031 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vctnx\" (UniqueName: \"kubernetes.io/projected/53784836-318f-4a01-899d-f3cd6cbd72d0-kube-api-access-vctnx\") pod \"watcher-kuttl-applier-0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.201751 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftxbv\" (UniqueName: \"kubernetes.io/projected/d4524e41-3177-4936-b995-6b0231394c16-kube-api-access-ftxbv\") pod \"watcher-kuttl-api-0\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.264397 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273175 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273245 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273310 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273374 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273414 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c847b0bd-b26d-4d23-a465-15d57f1a968e-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273446 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h8lz\" (UniqueName: \"kubernetes.io/projected/c847b0bd-b26d-4d23-a465-15d57f1a968e-kube-api-access-2h8lz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.273849 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c847b0bd-b26d-4d23-a465-15d57f1a968e-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.277289 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.277738 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.278022 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.278416 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.283285 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.290734 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h8lz\" (UniqueName: \"kubernetes.io/projected/c847b0bd-b26d-4d23-a465-15d57f1a968e-kube-api-access-2h8lz\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.362945 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.748191 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:53:31 crc kubenswrapper[4843]: W0314 09:53:31.757341 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4524e41_3177_4936_b995_6b0231394c16.slice/crio-d3de27c941bc73b9effafc1c2d7b8d8e6adf8b335c4542126549ede4772a3b57 WatchSource:0}: Error finding container d3de27c941bc73b9effafc1c2d7b8d8e6adf8b335c4542126549ede4772a3b57: Status 404 returned error can't find the container with id d3de27c941bc73b9effafc1c2d7b8d8e6adf8b335c4542126549ede4772a3b57 Mar 14 09:53:31 crc kubenswrapper[4843]: I0314 09:53:31.898589 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:53:31 crc kubenswrapper[4843]: W0314 09:53:31.921975 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53784836_318f_4a01_899d_f3cd6cbd72d0.slice/crio-e1edbecd83725c596f781b5d784089c89050e6b1b15f23a9e5214826ef76b2f3 WatchSource:0}: Error finding container e1edbecd83725c596f781b5d784089c89050e6b1b15f23a9e5214826ef76b2f3: Status 404 returned error can't find the container with id e1edbecd83725c596f781b5d784089c89050e6b1b15f23a9e5214826ef76b2f3 Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.182440 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:53:32 crc kubenswrapper[4843]: W0314 09:53:32.194429 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc847b0bd_b26d_4d23_a465_15d57f1a968e.slice/crio-be5e87ab6227d3a6088a6080a272ac1323895c4cadfb7134b0c1a67f14392a18 WatchSource:0}: Error finding container be5e87ab6227d3a6088a6080a272ac1323895c4cadfb7134b0c1a67f14392a18: Status 404 returned error can't find the container with id be5e87ab6227d3a6088a6080a272ac1323895c4cadfb7134b0c1a67f14392a18 Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.661680 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d4524e41-3177-4936-b995-6b0231394c16","Type":"ContainerStarted","Data":"efe65270a070c09c320a96842e53a64e21ee058c9ab97a7d1731b6135fc64e28"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.661722 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d4524e41-3177-4936-b995-6b0231394c16","Type":"ContainerStarted","Data":"244741a0a9f9acfe64beeb18c8adfd7376b0f149919d85748105035f225f3558"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.661734 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d4524e41-3177-4936-b995-6b0231394c16","Type":"ContainerStarted","Data":"d3de27c941bc73b9effafc1c2d7b8d8e6adf8b335c4542126549ede4772a3b57"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.661904 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.664175 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"53784836-318f-4a01-899d-f3cd6cbd72d0","Type":"ContainerStarted","Data":"53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.664229 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"53784836-318f-4a01-899d-f3cd6cbd72d0","Type":"ContainerStarted","Data":"e1edbecd83725c596f781b5d784089c89050e6b1b15f23a9e5214826ef76b2f3"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.665791 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"c847b0bd-b26d-4d23-a465-15d57f1a968e","Type":"ContainerStarted","Data":"6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.665830 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"c847b0bd-b26d-4d23-a465-15d57f1a968e","Type":"ContainerStarted","Data":"be5e87ab6227d3a6088a6080a272ac1323895c4cadfb7134b0c1a67f14392a18"} Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.685513 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.68549906 podStartE2EDuration="2.68549906s" podCreationTimestamp="2026-03-14 09:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:32.680707842 +0000 UTC m=+2579.993318980" watchObservedRunningTime="2026-03-14 09:53:32.68549906 +0000 UTC m=+2579.998110198" Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.702713 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=1.702693881 podStartE2EDuration="1.702693881s" podCreationTimestamp="2026-03-14 09:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:32.695326971 +0000 UTC m=+2580.007938099" watchObservedRunningTime="2026-03-14 09:53:32.702693881 +0000 UTC m=+2580.015305009" Mar 14 09:53:32 crc kubenswrapper[4843]: I0314 09:53:32.718578 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.718559849 podStartE2EDuration="2.718559849s" podCreationTimestamp="2026-03-14 09:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:32.717311768 +0000 UTC m=+2580.029922896" watchObservedRunningTime="2026-03-14 09:53:32.718559849 +0000 UTC m=+2580.031170977" Mar 14 09:53:33 crc kubenswrapper[4843]: I0314 09:53:33.090961 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:34 crc kubenswrapper[4843]: I0314 09:53:34.283532 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:35 crc kubenswrapper[4843]: I0314 09:53:35.037941 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:35 crc kubenswrapper[4843]: I0314 09:53:35.446440 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:36 crc kubenswrapper[4843]: I0314 09:53:36.264577 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:36 crc kubenswrapper[4843]: I0314 09:53:36.284419 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:36 crc kubenswrapper[4843]: I0314 09:53:36.614526 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:37 crc kubenswrapper[4843]: I0314 09:53:37.850518 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:39 crc kubenswrapper[4843]: I0314 09:53:39.113098 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:40 crc kubenswrapper[4843]: I0314 09:53:40.363752 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.264607 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.284982 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.290326 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.290435 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.363291 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.385584 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.605933 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.739442 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.744917 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.784073 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:53:41 crc kubenswrapper[4843]: I0314 09:53:41.786681 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:53:42 crc kubenswrapper[4843]: I0314 09:53:42.825751 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.078250 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.543426 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-db-create-kwwhw"] Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.544500 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.558089 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-create-kwwhw"] Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.572241 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-9191-account-create-update-hwvmk"] Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.573748 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.576212 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-9191-account-create-update-hwvmk"] Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.584800 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-db-secret" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.688547 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6zts\" (UniqueName: \"kubernetes.io/projected/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-kube-api-access-g6zts\") pod \"cinder-db-create-kwwhw\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.688607 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-operator-scripts\") pod \"cinder-db-create-kwwhw\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.688750 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8885224-4aed-4a40-8dfa-89b64366597f-operator-scripts\") pod \"cinder-9191-account-create-update-hwvmk\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.688799 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtkr2\" (UniqueName: \"kubernetes.io/projected/c8885224-4aed-4a40-8dfa-89b64366597f-kube-api-access-wtkr2\") pod \"cinder-9191-account-create-update-hwvmk\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.790153 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6zts\" (UniqueName: \"kubernetes.io/projected/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-kube-api-access-g6zts\") pod \"cinder-db-create-kwwhw\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.790213 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-operator-scripts\") pod \"cinder-db-create-kwwhw\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.790264 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8885224-4aed-4a40-8dfa-89b64366597f-operator-scripts\") pod \"cinder-9191-account-create-update-hwvmk\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.790309 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtkr2\" (UniqueName: \"kubernetes.io/projected/c8885224-4aed-4a40-8dfa-89b64366597f-kube-api-access-wtkr2\") pod \"cinder-9191-account-create-update-hwvmk\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.791314 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-operator-scripts\") pod \"cinder-db-create-kwwhw\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.791762 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8885224-4aed-4a40-8dfa-89b64366597f-operator-scripts\") pod \"cinder-9191-account-create-update-hwvmk\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.811205 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtkr2\" (UniqueName: \"kubernetes.io/projected/c8885224-4aed-4a40-8dfa-89b64366597f-kube-api-access-wtkr2\") pod \"cinder-9191-account-create-update-hwvmk\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.820042 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6zts\" (UniqueName: \"kubernetes.io/projected/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-kube-api-access-g6zts\") pod \"cinder-db-create-kwwhw\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.916148 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:43 crc kubenswrapper[4843]: I0314 09:53:43.925708 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.122923 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.123187 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-central-agent" containerID="cri-o://307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437" gracePeriod=30 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.124339 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="proxy-httpd" containerID="cri-o://f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae" gracePeriod=30 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.124720 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-notification-agent" containerID="cri-o://4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1" gracePeriod=30 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.124777 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="sg-core" containerID="cri-o://8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76" gracePeriod=30 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.317495 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.414245 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-9191-account-create-update-hwvmk"] Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.437794 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-create-kwwhw"] Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.504096 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.237:3000/\": read tcp 10.217.0.2:47026->10.217.0.237:3000: read: connection reset by peer" Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.766142 4843 generic.go:334] "Generic (PLEG): container finished" podID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerID="f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae" exitCode=0 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.766179 4843 generic.go:334] "Generic (PLEG): container finished" podID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerID="8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76" exitCode=2 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.766191 4843 generic.go:334] "Generic (PLEG): container finished" podID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerID="307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437" exitCode=0 Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.766235 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerDied","Data":"f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.766266 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerDied","Data":"8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.766302 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerDied","Data":"307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.767971 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-kwwhw" event={"ID":"60eeb6e3-bcc2-4c90-b551-0db727abcb1d","Type":"ContainerStarted","Data":"8e1c164eba9a345cd73527bb68de74431c2d41d2aa0373385ab4bd32751b3bb3"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.768006 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-kwwhw" event={"ID":"60eeb6e3-bcc2-4c90-b551-0db727abcb1d","Type":"ContainerStarted","Data":"8d00acd7dbfc1ed8804c51fd782a9a72668504f78ed4e9a4742ad79b6ee99032"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.771982 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" event={"ID":"c8885224-4aed-4a40-8dfa-89b64366597f","Type":"ContainerStarted","Data":"fad3d5e4665168053df9e867b63377967b5b1a4e96396a3e1e54aac815fca711"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.772028 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" event={"ID":"c8885224-4aed-4a40-8dfa-89b64366597f","Type":"ContainerStarted","Data":"580b71cbdd290b46d4b9dfc32b5e813eb2d555f987914b3b24918f13b3c07d8d"} Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.794660 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-db-create-kwwhw" podStartSLOduration=1.794635595 podStartE2EDuration="1.794635595s" podCreationTimestamp="2026-03-14 09:53:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:44.790089024 +0000 UTC m=+2592.102700152" watchObservedRunningTime="2026-03-14 09:53:44.794635595 +0000 UTC m=+2592.107246733" Mar 14 09:53:44 crc kubenswrapper[4843]: I0314 09:53:44.809542 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" podStartSLOduration=1.8095251399999999 podStartE2EDuration="1.80952514s" podCreationTimestamp="2026-03-14 09:53:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:53:44.803975074 +0000 UTC m=+2592.116586222" watchObservedRunningTime="2026-03-14 09:53:44.80952514 +0000 UTC m=+2592.122136288" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.563911 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.610175 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.731859 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-scripts\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.731996 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q8w2\" (UniqueName: \"kubernetes.io/projected/42c05d98-1fe3-4a70-ac34-8f3601433809-kube-api-access-6q8w2\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.732846 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-run-httpd\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.732898 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-config-data\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.732937 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-sg-core-conf-yaml\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.732985 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-combined-ca-bundle\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.733058 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-ceilometer-tls-certs\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.733095 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-log-httpd\") pod \"42c05d98-1fe3-4a70-ac34-8f3601433809\" (UID: \"42c05d98-1fe3-4a70-ac34-8f3601433809\") " Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.733256 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.733751 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.734082 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.737761 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42c05d98-1fe3-4a70-ac34-8f3601433809-kube-api-access-6q8w2" (OuterVolumeSpecName: "kube-api-access-6q8w2") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "kube-api-access-6q8w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.748265 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-scripts" (OuterVolumeSpecName: "scripts") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.757192 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.786823 4843 generic.go:334] "Generic (PLEG): container finished" podID="c8885224-4aed-4a40-8dfa-89b64366597f" containerID="fad3d5e4665168053df9e867b63377967b5b1a4e96396a3e1e54aac815fca711" exitCode=0 Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.786896 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" event={"ID":"c8885224-4aed-4a40-8dfa-89b64366597f","Type":"ContainerDied","Data":"fad3d5e4665168053df9e867b63377967b5b1a4e96396a3e1e54aac815fca711"} Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.793636 4843 generic.go:334] "Generic (PLEG): container finished" podID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerID="4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1" exitCode=0 Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.793793 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerDied","Data":"4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1"} Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.793823 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"42c05d98-1fe3-4a70-ac34-8f3601433809","Type":"ContainerDied","Data":"bc9dc11f44cb1ab237cd6dd5450c1c5da73437461af281f9e7ea23b73d9d4236"} Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.793860 4843 scope.go:117] "RemoveContainer" containerID="f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.794085 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.797221 4843 generic.go:334] "Generic (PLEG): container finished" podID="60eeb6e3-bcc2-4c90-b551-0db727abcb1d" containerID="8e1c164eba9a345cd73527bb68de74431c2d41d2aa0373385ab4bd32751b3bb3" exitCode=0 Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.797262 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-kwwhw" event={"ID":"60eeb6e3-bcc2-4c90-b551-0db727abcb1d","Type":"ContainerDied","Data":"8e1c164eba9a345cd73527bb68de74431c2d41d2aa0373385ab4bd32751b3bb3"} Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.814453 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.816640 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.835202 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.835228 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q8w2\" (UniqueName: \"kubernetes.io/projected/42c05d98-1fe3-4a70-ac34-8f3601433809-kube-api-access-6q8w2\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.835239 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.835247 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.835254 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.835262 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42c05d98-1fe3-4a70-ac34-8f3601433809-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.841005 4843 scope.go:117] "RemoveContainer" containerID="8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.854561 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-config-data" (OuterVolumeSpecName: "config-data") pod "42c05d98-1fe3-4a70-ac34-8f3601433809" (UID: "42c05d98-1fe3-4a70-ac34-8f3601433809"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.925884 4843 scope.go:117] "RemoveContainer" containerID="4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.936758 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42c05d98-1fe3-4a70-ac34-8f3601433809-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.941260 4843 scope.go:117] "RemoveContainer" containerID="307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.960745 4843 scope.go:117] "RemoveContainer" containerID="f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae" Mar 14 09:53:45 crc kubenswrapper[4843]: E0314 09:53:45.961133 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae\": container with ID starting with f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae not found: ID does not exist" containerID="f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.961163 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae"} err="failed to get container status \"f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae\": rpc error: code = NotFound desc = could not find container \"f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae\": container with ID starting with f621380868a24c7dd4e97c010b20d4088d89f3f42601b5e51751f04d16daffae not found: ID does not exist" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.961186 4843 scope.go:117] "RemoveContainer" containerID="8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76" Mar 14 09:53:45 crc kubenswrapper[4843]: E0314 09:53:45.961420 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76\": container with ID starting with 8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76 not found: ID does not exist" containerID="8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.961509 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76"} err="failed to get container status \"8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76\": rpc error: code = NotFound desc = could not find container \"8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76\": container with ID starting with 8a3b2a9ff08e68ea9aaf6f2514201ac88210db1569fe417228800eba4a41ad76 not found: ID does not exist" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.961571 4843 scope.go:117] "RemoveContainer" containerID="4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1" Mar 14 09:53:45 crc kubenswrapper[4843]: E0314 09:53:45.962355 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1\": container with ID starting with 4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1 not found: ID does not exist" containerID="4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.962396 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1"} err="failed to get container status \"4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1\": rpc error: code = NotFound desc = could not find container \"4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1\": container with ID starting with 4980234ffb7a8bf0ee9aa23b7c9d15106050a835b7155c620f5d8f5f602145a1 not found: ID does not exist" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.962429 4843 scope.go:117] "RemoveContainer" containerID="307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437" Mar 14 09:53:45 crc kubenswrapper[4843]: E0314 09:53:45.962697 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437\": container with ID starting with 307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437 not found: ID does not exist" containerID="307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437" Mar 14 09:53:45 crc kubenswrapper[4843]: I0314 09:53:45.962731 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437"} err="failed to get container status \"307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437\": rpc error: code = NotFound desc = could not find container \"307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437\": container with ID starting with 307fd4b7e21c8d614fa58353f5a6fe1a555d0dec7e624e58b50a816f6e0cd437 not found: ID does not exist" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.124846 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.132501 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.150488 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:46 crc kubenswrapper[4843]: E0314 09:53:46.150859 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="sg-core" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.150879 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="sg-core" Mar 14 09:53:46 crc kubenswrapper[4843]: E0314 09:53:46.150897 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="proxy-httpd" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.150905 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="proxy-httpd" Mar 14 09:53:46 crc kubenswrapper[4843]: E0314 09:53:46.150929 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-notification-agent" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.150937 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-notification-agent" Mar 14 09:53:46 crc kubenswrapper[4843]: E0314 09:53:46.150951 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-central-agent" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.150959 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-central-agent" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.151160 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-notification-agent" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.151200 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="proxy-httpd" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.151213 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="ceilometer-central-agent" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.151223 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" containerName="sg-core" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.153249 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.155098 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.155316 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.155320 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.166964 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239002 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239069 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-config-data\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239086 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239105 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-log-httpd\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239120 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhc7p\" (UniqueName: \"kubernetes.io/projected/6cce67ea-d82a-441b-a52b-eb025169a40b-kube-api-access-mhc7p\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239141 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239185 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-run-httpd\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.239219 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-scripts\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.340664 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-scripts\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.340847 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.340963 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-config-data\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.340982 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.341002 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-log-httpd\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.341016 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhc7p\" (UniqueName: \"kubernetes.io/projected/6cce67ea-d82a-441b-a52b-eb025169a40b-kube-api-access-mhc7p\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.341059 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.341077 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-run-httpd\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.341796 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-log-httpd\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.343083 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-run-httpd\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.345578 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-config-data\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.345792 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-scripts\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.345798 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.346145 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.349853 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.358912 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhc7p\" (UniqueName: \"kubernetes.io/projected/6cce67ea-d82a-441b-a52b-eb025169a40b-kube-api-access-mhc7p\") pod \"ceilometer-0\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.473447 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.770946 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:46 crc kubenswrapper[4843]: I0314 09:53:46.943589 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:53:46 crc kubenswrapper[4843]: W0314 09:53:46.951472 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cce67ea_d82a_441b_a52b_eb025169a40b.slice/crio-2fd1df63b051ff30d4e09eaf26e80cdf9d238d6d701b0178eebd11cefe0dc331 WatchSource:0}: Error finding container 2fd1df63b051ff30d4e09eaf26e80cdf9d238d6d701b0178eebd11cefe0dc331: Status 404 returned error can't find the container with id 2fd1df63b051ff30d4e09eaf26e80cdf9d238d6d701b0178eebd11cefe0dc331 Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.289985 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.293862 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.352496 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42c05d98-1fe3-4a70-ac34-8f3601433809" path="/var/lib/kubelet/pods/42c05d98-1fe3-4a70-ac34-8f3601433809/volumes" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.366630 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtkr2\" (UniqueName: \"kubernetes.io/projected/c8885224-4aed-4a40-8dfa-89b64366597f-kube-api-access-wtkr2\") pod \"c8885224-4aed-4a40-8dfa-89b64366597f\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.366700 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6zts\" (UniqueName: \"kubernetes.io/projected/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-kube-api-access-g6zts\") pod \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.366748 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-operator-scripts\") pod \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\" (UID: \"60eeb6e3-bcc2-4c90-b551-0db727abcb1d\") " Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.366785 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8885224-4aed-4a40-8dfa-89b64366597f-operator-scripts\") pod \"c8885224-4aed-4a40-8dfa-89b64366597f\" (UID: \"c8885224-4aed-4a40-8dfa-89b64366597f\") " Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.367798 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60eeb6e3-bcc2-4c90-b551-0db727abcb1d" (UID: "60eeb6e3-bcc2-4c90-b551-0db727abcb1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.367828 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8885224-4aed-4a40-8dfa-89b64366597f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8885224-4aed-4a40-8dfa-89b64366597f" (UID: "c8885224-4aed-4a40-8dfa-89b64366597f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.372377 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-kube-api-access-g6zts" (OuterVolumeSpecName: "kube-api-access-g6zts") pod "60eeb6e3-bcc2-4c90-b551-0db727abcb1d" (UID: "60eeb6e3-bcc2-4c90-b551-0db727abcb1d"). InnerVolumeSpecName "kube-api-access-g6zts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.372651 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8885224-4aed-4a40-8dfa-89b64366597f-kube-api-access-wtkr2" (OuterVolumeSpecName: "kube-api-access-wtkr2") pod "c8885224-4aed-4a40-8dfa-89b64366597f" (UID: "c8885224-4aed-4a40-8dfa-89b64366597f"). InnerVolumeSpecName "kube-api-access-wtkr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.468321 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtkr2\" (UniqueName: \"kubernetes.io/projected/c8885224-4aed-4a40-8dfa-89b64366597f-kube-api-access-wtkr2\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.468350 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6zts\" (UniqueName: \"kubernetes.io/projected/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-kube-api-access-g6zts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.468376 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60eeb6e3-bcc2-4c90-b551-0db727abcb1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.468385 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8885224-4aed-4a40-8dfa-89b64366597f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.825898 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-create-kwwhw" event={"ID":"60eeb6e3-bcc2-4c90-b551-0db727abcb1d","Type":"ContainerDied","Data":"8d00acd7dbfc1ed8804c51fd782a9a72668504f78ed4e9a4742ad79b6ee99032"} Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.825944 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d00acd7dbfc1ed8804c51fd782a9a72668504f78ed4e9a4742ad79b6ee99032" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.825969 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-create-kwwhw" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.827704 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.827746 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-9191-account-create-update-hwvmk" event={"ID":"c8885224-4aed-4a40-8dfa-89b64366597f","Type":"ContainerDied","Data":"580b71cbdd290b46d4b9dfc32b5e813eb2d555f987914b3b24918f13b3c07d8d"} Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.828110 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="580b71cbdd290b46d4b9dfc32b5e813eb2d555f987914b3b24918f13b3c07d8d" Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.829702 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerStarted","Data":"8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b"} Mar 14 09:53:47 crc kubenswrapper[4843]: I0314 09:53:47.829741 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerStarted","Data":"2fd1df63b051ff30d4e09eaf26e80cdf9d238d6d701b0178eebd11cefe0dc331"} Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.015679 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.846338 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerStarted","Data":"24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31"} Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.846613 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerStarted","Data":"11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822"} Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.887050 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-db-sync-9kdvr"] Mar 14 09:53:48 crc kubenswrapper[4843]: E0314 09:53:48.887417 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60eeb6e3-bcc2-4c90-b551-0db727abcb1d" containerName="mariadb-database-create" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.887436 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="60eeb6e3-bcc2-4c90-b551-0db727abcb1d" containerName="mariadb-database-create" Mar 14 09:53:48 crc kubenswrapper[4843]: E0314 09:53:48.887450 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8885224-4aed-4a40-8dfa-89b64366597f" containerName="mariadb-account-create-update" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.887457 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8885224-4aed-4a40-8dfa-89b64366597f" containerName="mariadb-account-create-update" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.887638 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8885224-4aed-4a40-8dfa-89b64366597f" containerName="mariadb-account-create-update" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.887661 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="60eeb6e3-bcc2-4c90-b551-0db727abcb1d" containerName="mariadb-database-create" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.888312 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.891751 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-config-data" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.891870 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scripts" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.891983 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-cinder-dockercfg-bb8mv" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.894681 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-9kdvr"] Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.995355 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-db-sync-config-data\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.995665 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2drq\" (UniqueName: \"kubernetes.io/projected/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-kube-api-access-n2drq\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.995716 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-config-data\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.995745 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-scripts\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.995838 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-etc-machine-id\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:48 crc kubenswrapper[4843]: I0314 09:53:48.995885 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-combined-ca-bundle\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.096756 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-combined-ca-bundle\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.096806 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-db-sync-config-data\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.096843 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2drq\" (UniqueName: \"kubernetes.io/projected/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-kube-api-access-n2drq\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.096884 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-config-data\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.096910 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-scripts\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.096951 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-etc-machine-id\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.097037 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-etc-machine-id\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.107191 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-config-data\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.113145 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-combined-ca-bundle\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.113948 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2drq\" (UniqueName: \"kubernetes.io/projected/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-kube-api-access-n2drq\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.121153 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-scripts\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.121697 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-db-sync-config-data\") pod \"cinder-db-sync-9kdvr\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.206367 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.221466 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.675689 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-9kdvr"] Mar 14 09:53:49 crc kubenswrapper[4843]: W0314 09:53:49.679368 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacaf6bc4_ecac_43b6_83c3_3509cf5a3fe6.slice/crio-8e98b9109dc0fc6779fb804a90f8675d32173fae25d401f60c19f3c428b472ba WatchSource:0}: Error finding container 8e98b9109dc0fc6779fb804a90f8675d32173fae25d401f60c19f3c428b472ba: Status 404 returned error can't find the container with id 8e98b9109dc0fc6779fb804a90f8675d32173fae25d401f60c19f3c428b472ba Mar 14 09:53:49 crc kubenswrapper[4843]: I0314 09:53:49.853869 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" event={"ID":"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6","Type":"ContainerStarted","Data":"8e98b9109dc0fc6779fb804a90f8675d32173fae25d401f60c19f3c428b472ba"} Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.403006 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.757757 4843 scope.go:117] "RemoveContainer" containerID="163b59a8f00ed4d144a7d383210261c8b1ddbc5e0ba3122195dce7602d3e6de2" Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.783744 4843 scope.go:117] "RemoveContainer" containerID="741291ee1cf83addf57cdd65999f6388854c5c20b1503a2e0c56d4de682f450c" Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.815889 4843 scope.go:117] "RemoveContainer" containerID="8124c7ce40ba05f5165d0a40f17fcda6f92cdc98efb114f09bfe50443426a48a" Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.867382 4843 scope.go:117] "RemoveContainer" containerID="2a985a0a7d0935235b6fc177d8b636f30d439f1568a8d03efafeb33918bae25c" Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.875347 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerStarted","Data":"f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a"} Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.875415 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:53:50 crc kubenswrapper[4843]: I0314 09:53:50.899214 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.73103042 podStartE2EDuration="4.899196681s" podCreationTimestamp="2026-03-14 09:53:46 +0000 UTC" firstStartedPulling="2026-03-14 09:53:46.955025627 +0000 UTC m=+2594.267636755" lastFinishedPulling="2026-03-14 09:53:50.123191878 +0000 UTC m=+2597.435803016" observedRunningTime="2026-03-14 09:53:50.89553105 +0000 UTC m=+2598.208142178" watchObservedRunningTime="2026-03-14 09:53:50.899196681 +0000 UTC m=+2598.211807809" Mar 14 09:53:51 crc kubenswrapper[4843]: I0314 09:53:51.688444 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:52 crc kubenswrapper[4843]: I0314 09:53:52.928324 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:54 crc kubenswrapper[4843]: I0314 09:53:54.176694 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:55 crc kubenswrapper[4843]: I0314 09:53:55.411810 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:56 crc kubenswrapper[4843]: I0314 09:53:56.685821 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:57 crc kubenswrapper[4843]: I0314 09:53:57.931960 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:53:59 crc kubenswrapper[4843]: I0314 09:53:59.212187 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.139573 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558034-cf7wr"] Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.143511 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.146486 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.147356 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.147588 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.158785 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558034-cf7wr"] Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.297850 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89wsl\" (UniqueName: \"kubernetes.io/projected/899a1e8a-4cee-4187-9d37-9afcc5c310db-kube-api-access-89wsl\") pod \"auto-csr-approver-29558034-cf7wr\" (UID: \"899a1e8a-4cee-4187-9d37-9afcc5c310db\") " pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.399764 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89wsl\" (UniqueName: \"kubernetes.io/projected/899a1e8a-4cee-4187-9d37-9afcc5c310db-kube-api-access-89wsl\") pod \"auto-csr-approver-29558034-cf7wr\" (UID: \"899a1e8a-4cee-4187-9d37-9afcc5c310db\") " pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.417520 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89wsl\" (UniqueName: \"kubernetes.io/projected/899a1e8a-4cee-4187-9d37-9afcc5c310db-kube-api-access-89wsl\") pod \"auto-csr-approver-29558034-cf7wr\" (UID: \"899a1e8a-4cee-4187-9d37-9afcc5c310db\") " pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.466298 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:00 crc kubenswrapper[4843]: I0314 09:54:00.472624 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:01 crc kubenswrapper[4843]: I0314 09:54:01.735004 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:02 crc kubenswrapper[4843]: I0314 09:54:02.929379 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:04 crc kubenswrapper[4843]: I0314 09:54:04.115230 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:05 crc kubenswrapper[4843]: I0314 09:54:05.352435 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:05 crc kubenswrapper[4843]: E0314 09:54:05.667040 4843 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 14 09:54:05 crc kubenswrapper[4843]: E0314 09:54:05.667499 4843 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n2drq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-9kdvr_watcher-kuttl-default(acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 14 09:54:05 crc kubenswrapper[4843]: E0314 09:54:05.669037 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" podUID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" Mar 14 09:54:06 crc kubenswrapper[4843]: I0314 09:54:06.051963 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558034-cf7wr"] Mar 14 09:54:06 crc kubenswrapper[4843]: W0314 09:54:06.059119 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod899a1e8a_4cee_4187_9d37_9afcc5c310db.slice/crio-8421642ea73a1cb113635754576f550207a5505cbdc55f86ba2fce601800e758 WatchSource:0}: Error finding container 8421642ea73a1cb113635754576f550207a5505cbdc55f86ba2fce601800e758: Status 404 returned error can't find the container with id 8421642ea73a1cb113635754576f550207a5505cbdc55f86ba2fce601800e758 Mar 14 09:54:06 crc kubenswrapper[4843]: I0314 09:54:06.115788 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" event={"ID":"899a1e8a-4cee-4187-9d37-9afcc5c310db","Type":"ContainerStarted","Data":"8421642ea73a1cb113635754576f550207a5505cbdc55f86ba2fce601800e758"} Mar 14 09:54:06 crc kubenswrapper[4843]: E0314 09:54:06.118065 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" podUID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" Mar 14 09:54:06 crc kubenswrapper[4843]: I0314 09:54:06.594715 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:07 crc kubenswrapper[4843]: I0314 09:54:07.817353 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:08 crc kubenswrapper[4843]: I0314 09:54:08.133080 4843 generic.go:334] "Generic (PLEG): container finished" podID="899a1e8a-4cee-4187-9d37-9afcc5c310db" containerID="6e944e23ca4b716eb4ca56bdc9761b797b2ed9c7df579a570eed58ad6368fe95" exitCode=0 Mar 14 09:54:08 crc kubenswrapper[4843]: I0314 09:54:08.133176 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" event={"ID":"899a1e8a-4cee-4187-9d37-9afcc5c310db","Type":"ContainerDied","Data":"6e944e23ca4b716eb4ca56bdc9761b797b2ed9c7df579a570eed58ad6368fe95"} Mar 14 09:54:09 crc kubenswrapper[4843]: I0314 09:54:09.019365 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:09 crc kubenswrapper[4843]: I0314 09:54:09.473674 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:09 crc kubenswrapper[4843]: I0314 09:54:09.593226 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89wsl\" (UniqueName: \"kubernetes.io/projected/899a1e8a-4cee-4187-9d37-9afcc5c310db-kube-api-access-89wsl\") pod \"899a1e8a-4cee-4187-9d37-9afcc5c310db\" (UID: \"899a1e8a-4cee-4187-9d37-9afcc5c310db\") " Mar 14 09:54:09 crc kubenswrapper[4843]: I0314 09:54:09.599894 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/899a1e8a-4cee-4187-9d37-9afcc5c310db-kube-api-access-89wsl" (OuterVolumeSpecName: "kube-api-access-89wsl") pod "899a1e8a-4cee-4187-9d37-9afcc5c310db" (UID: "899a1e8a-4cee-4187-9d37-9afcc5c310db"). InnerVolumeSpecName "kube-api-access-89wsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:09 crc kubenswrapper[4843]: I0314 09:54:09.695576 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89wsl\" (UniqueName: \"kubernetes.io/projected/899a1e8a-4cee-4187-9d37-9afcc5c310db-kube-api-access-89wsl\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:10 crc kubenswrapper[4843]: I0314 09:54:10.157358 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" event={"ID":"899a1e8a-4cee-4187-9d37-9afcc5c310db","Type":"ContainerDied","Data":"8421642ea73a1cb113635754576f550207a5505cbdc55f86ba2fce601800e758"} Mar 14 09:54:10 crc kubenswrapper[4843]: I0314 09:54:10.157398 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8421642ea73a1cb113635754576f550207a5505cbdc55f86ba2fce601800e758" Mar 14 09:54:10 crc kubenswrapper[4843]: I0314 09:54:10.157423 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558034-cf7wr" Mar 14 09:54:10 crc kubenswrapper[4843]: I0314 09:54:10.280510 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:10 crc kubenswrapper[4843]: I0314 09:54:10.549438 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558028-b8mw2"] Mar 14 09:54:10 crc kubenswrapper[4843]: I0314 09:54:10.563889 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558028-b8mw2"] Mar 14 09:54:11 crc kubenswrapper[4843]: I0314 09:54:11.351505 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b" path="/var/lib/kubelet/pods/7f0f2ca8-7ffe-4d0f-8138-b6e25849a29b/volumes" Mar 14 09:54:11 crc kubenswrapper[4843]: I0314 09:54:11.539724 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:12 crc kubenswrapper[4843]: I0314 09:54:12.799399 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:14 crc kubenswrapper[4843]: I0314 09:54:14.021061 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:15 crc kubenswrapper[4843]: I0314 09:54:15.226835 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:16 crc kubenswrapper[4843]: I0314 09:54:16.438166 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:16 crc kubenswrapper[4843]: I0314 09:54:16.482453 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:17 crc kubenswrapper[4843]: I0314 09:54:17.631733 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:18 crc kubenswrapper[4843]: I0314 09:54:18.884623 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:20 crc kubenswrapper[4843]: I0314 09:54:20.139907 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:21 crc kubenswrapper[4843]: I0314 09:54:21.256781 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" event={"ID":"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6","Type":"ContainerStarted","Data":"94e5e55eb4e83ba29e712badf5917865efd660c2b0fd04d836960d5e61d37015"} Mar 14 09:54:21 crc kubenswrapper[4843]: I0314 09:54:21.283603 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" podStartSLOduration=3.197984618 podStartE2EDuration="33.283580321s" podCreationTimestamp="2026-03-14 09:53:48 +0000 UTC" firstStartedPulling="2026-03-14 09:53:49.681368709 +0000 UTC m=+2596.993979837" lastFinishedPulling="2026-03-14 09:54:19.766964372 +0000 UTC m=+2627.079575540" observedRunningTime="2026-03-14 09:54:21.281331576 +0000 UTC m=+2628.593942754" watchObservedRunningTime="2026-03-14 09:54:21.283580321 +0000 UTC m=+2628.596191479" Mar 14 09:54:21 crc kubenswrapper[4843]: I0314 09:54:21.412486 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:22 crc kubenswrapper[4843]: I0314 09:54:22.634987 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:23 crc kubenswrapper[4843]: I0314 09:54:23.843471 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:24 crc kubenswrapper[4843]: I0314 09:54:24.285324 4843 generic.go:334] "Generic (PLEG): container finished" podID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" containerID="94e5e55eb4e83ba29e712badf5917865efd660c2b0fd04d836960d5e61d37015" exitCode=0 Mar 14 09:54:24 crc kubenswrapper[4843]: I0314 09:54:24.285390 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" event={"ID":"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6","Type":"ContainerDied","Data":"94e5e55eb4e83ba29e712badf5917865efd660c2b0fd04d836960d5e61d37015"} Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.112352 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.675494 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693547 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-scripts\") pod \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693606 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-combined-ca-bundle\") pod \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693676 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-config-data\") pod \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693704 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2drq\" (UniqueName: \"kubernetes.io/projected/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-kube-api-access-n2drq\") pod \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693729 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-db-sync-config-data\") pod \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693769 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-etc-machine-id\") pod \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\" (UID: \"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6\") " Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.693993 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" (UID: "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.698684 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" (UID: "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.699663 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-kube-api-access-n2drq" (OuterVolumeSpecName: "kube-api-access-n2drq") pod "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" (UID: "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6"). InnerVolumeSpecName "kube-api-access-n2drq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.702008 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-scripts" (OuterVolumeSpecName: "scripts") pod "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" (UID: "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.718508 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" (UID: "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.750583 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-config-data" (OuterVolumeSpecName: "config-data") pod "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" (UID: "acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.794955 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2drq\" (UniqueName: \"kubernetes.io/projected/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-kube-api-access-n2drq\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.794988 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.794997 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.795007 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.795016 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:25 crc kubenswrapper[4843]: I0314 09:54:25.795025 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.325401 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.326298 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" event={"ID":"acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6","Type":"ContainerDied","Data":"8e98b9109dc0fc6779fb804a90f8675d32173fae25d401f60c19f3c428b472ba"} Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.326330 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e98b9109dc0fc6779fb804a90f8675d32173fae25d401f60c19f3c428b472ba" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.326407 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-db-sync-9kdvr" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.591454 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:26 crc kubenswrapper[4843]: E0314 09:54:26.591888 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899a1e8a-4cee-4187-9d37-9afcc5c310db" containerName="oc" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.591903 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="899a1e8a-4cee-4187-9d37-9afcc5c310db" containerName="oc" Mar 14 09:54:26 crc kubenswrapper[4843]: E0314 09:54:26.591946 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" containerName="cinder-db-sync" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.591955 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" containerName="cinder-db-sync" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.592122 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" containerName="cinder-db-sync" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.592138 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="899a1e8a-4cee-4187-9d37-9afcc5c310db" containerName="oc" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.593217 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.602020 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scheduler-config-data" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.605835 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-cinder-dockercfg-bb8mv" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.608897 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-config-data" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.613698 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scripts" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.635794 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.637684 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.640824 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-backup-config-data" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.657337 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.680214 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712352 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snrbq\" (UniqueName: \"kubernetes.io/projected/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-kube-api-access-snrbq\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712421 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712497 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712566 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712625 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712642 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-scripts\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.712669 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.813692 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.813760 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.813813 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.813834 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.813908 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.813969 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814021 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-scripts\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814042 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814108 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814174 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814234 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdvft\" (UniqueName: \"kubernetes.io/projected/d220fd8b-4d05-498c-8a78-7035f12b054c-kube-api-access-cdvft\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814343 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-lib-modules\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814455 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814502 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814545 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814567 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-scripts\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814607 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814654 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814722 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snrbq\" (UniqueName: \"kubernetes.io/projected/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-kube-api-access-snrbq\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814759 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-sys\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814790 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-run\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814825 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814851 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-dev\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.814944 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.823831 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.824427 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.824960 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.825960 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.827707 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-scripts\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.848262 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snrbq\" (UniqueName: \"kubernetes.io/projected/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-kube-api-access-snrbq\") pod \"cinder-scheduler-0\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918166 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918255 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918312 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdvft\" (UniqueName: \"kubernetes.io/projected/d220fd8b-4d05-498c-8a78-7035f12b054c-kube-api-access-cdvft\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918334 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-lib-modules\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918366 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918395 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918421 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918463 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-sys\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918483 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-run\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918512 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-dev\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918544 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918579 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918600 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918629 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918658 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.918681 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-scripts\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.919393 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-sys\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.919636 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.923464 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.923539 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-run\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.923574 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-dev\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.923609 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.923966 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-lib-modules\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.924013 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.924926 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.925767 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.926047 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.929550 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.930016 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-scripts\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.931011 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.937045 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.940010 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.971904 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdvft\" (UniqueName: \"kubernetes.io/projected/d220fd8b-4d05-498c-8a78-7035f12b054c-kube-api-access-cdvft\") pod \"cinder-backup-0\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:26 crc kubenswrapper[4843]: I0314 09:54:26.974640 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.011887 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.013189 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.028873 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-api-config-data" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.033582 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125191 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125244 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38083617-1609-475c-98dc-f475249194bd-logs\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125294 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125318 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38083617-1609-475c-98dc-f475249194bd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125353 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data-custom\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125378 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfr77\" (UniqueName: \"kubernetes.io/projected/38083617-1609-475c-98dc-f475249194bd-kube-api-access-zfr77\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125401 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-scripts\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.125435 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.226344 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfr77\" (UniqueName: \"kubernetes.io/projected/38083617-1609-475c-98dc-f475249194bd-kube-api-access-zfr77\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.226581 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-scripts\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.226623 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.226682 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.226705 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38083617-1609-475c-98dc-f475249194bd-logs\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.226739 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.227237 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38083617-1609-475c-98dc-f475249194bd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.227299 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data-custom\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.228301 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38083617-1609-475c-98dc-f475249194bd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.229238 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38083617-1609-475c-98dc-f475249194bd-logs\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.233033 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-scripts\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.233687 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.235427 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data-custom\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.236189 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.240339 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.255194 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfr77\" (UniqueName: \"kubernetes.io/projected/38083617-1609-475c-98dc-f475249194bd-kube-api-access-zfr77\") pod \"cinder-api-0\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.414641 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.488002 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.612318 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.708546 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:27 crc kubenswrapper[4843]: I0314 09:54:27.877421 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:27 crc kubenswrapper[4843]: W0314 09:54:27.882385 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38083617_1609_475c_98dc_f475249194bd.slice/crio-f4798bff6edc25d2f8ec3ab859e1cb7dcd9985f1e7dc21f3d89642c5a19ff854 WatchSource:0}: Error finding container f4798bff6edc25d2f8ec3ab859e1cb7dcd9985f1e7dc21f3d89642c5a19ff854: Status 404 returned error can't find the container with id f4798bff6edc25d2f8ec3ab859e1cb7dcd9985f1e7dc21f3d89642c5a19ff854 Mar 14 09:54:28 crc kubenswrapper[4843]: I0314 09:54:28.361099 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"d220fd8b-4d05-498c-8a78-7035f12b054c","Type":"ContainerStarted","Data":"8ae46251efa386aac6be482e0e5083b623fab593e69cb74181a435588374faf0"} Mar 14 09:54:28 crc kubenswrapper[4843]: I0314 09:54:28.369952 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"38083617-1609-475c-98dc-f475249194bd","Type":"ContainerStarted","Data":"f4798bff6edc25d2f8ec3ab859e1cb7dcd9985f1e7dc21f3d89642c5a19ff854"} Mar 14 09:54:28 crc kubenswrapper[4843]: I0314 09:54:28.371571 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090","Type":"ContainerStarted","Data":"7e68db2d8324f6939668ec1c9911a9f9aacfc34b93d67213f9defb4b09eba8cd"} Mar 14 09:54:28 crc kubenswrapper[4843]: I0314 09:54:28.735545 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:29 crc kubenswrapper[4843]: I0314 09:54:29.263208 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:29 crc kubenswrapper[4843]: I0314 09:54:29.493005 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"d220fd8b-4d05-498c-8a78-7035f12b054c","Type":"ContainerStarted","Data":"c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5"} Mar 14 09:54:29 crc kubenswrapper[4843]: I0314 09:54:29.493077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"d220fd8b-4d05-498c-8a78-7035f12b054c","Type":"ContainerStarted","Data":"731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d"} Mar 14 09:54:29 crc kubenswrapper[4843]: I0314 09:54:29.504046 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"38083617-1609-475c-98dc-f475249194bd","Type":"ContainerStarted","Data":"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0"} Mar 14 09:54:29 crc kubenswrapper[4843]: I0314 09:54:29.529379 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-backup-0" podStartSLOduration=2.688730636 podStartE2EDuration="3.529356641s" podCreationTimestamp="2026-03-14 09:54:26 +0000 UTC" firstStartedPulling="2026-03-14 09:54:27.625811848 +0000 UTC m=+2634.938422976" lastFinishedPulling="2026-03-14 09:54:28.466437853 +0000 UTC m=+2635.779048981" observedRunningTime="2026-03-14 09:54:29.528298055 +0000 UTC m=+2636.840909183" watchObservedRunningTime="2026-03-14 09:54:29.529356641 +0000 UTC m=+2636.841967769" Mar 14 09:54:29 crc kubenswrapper[4843]: I0314 09:54:29.928173 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:30 crc kubenswrapper[4843]: I0314 09:54:30.515227 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"38083617-1609-475c-98dc-f475249194bd","Type":"ContainerStarted","Data":"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917"} Mar 14 09:54:30 crc kubenswrapper[4843]: I0314 09:54:30.515514 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:30 crc kubenswrapper[4843]: I0314 09:54:30.515409 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api" containerID="cri-o://d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917" gracePeriod=30 Mar 14 09:54:30 crc kubenswrapper[4843]: I0314 09:54:30.515331 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api-log" containerID="cri-o://cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0" gracePeriod=30 Mar 14 09:54:30 crc kubenswrapper[4843]: I0314 09:54:30.519666 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090","Type":"ContainerStarted","Data":"177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8"} Mar 14 09:54:30 crc kubenswrapper[4843]: I0314 09:54:30.545526 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-api-0" podStartSLOduration=4.545507684 podStartE2EDuration="4.545507684s" podCreationTimestamp="2026-03-14 09:54:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:54:30.541781092 +0000 UTC m=+2637.854392220" watchObservedRunningTime="2026-03-14 09:54:30.545507684 +0000 UTC m=+2637.858118802" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.164648 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.520125 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529255 4843 generic.go:334] "Generic (PLEG): container finished" podID="38083617-1609-475c-98dc-f475249194bd" containerID="d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917" exitCode=0 Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529295 4843 generic.go:334] "Generic (PLEG): container finished" podID="38083617-1609-475c-98dc-f475249194bd" containerID="cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0" exitCode=143 Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529331 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"38083617-1609-475c-98dc-f475249194bd","Type":"ContainerDied","Data":"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917"} Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529357 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"38083617-1609-475c-98dc-f475249194bd","Type":"ContainerDied","Data":"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0"} Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529368 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"38083617-1609-475c-98dc-f475249194bd","Type":"ContainerDied","Data":"f4798bff6edc25d2f8ec3ab859e1cb7dcd9985f1e7dc21f3d89642c5a19ff854"} Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529383 4843 scope.go:117] "RemoveContainer" containerID="d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.529406 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.532420 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090","Type":"ContainerStarted","Data":"6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f"} Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.552406 4843 scope.go:117] "RemoveContainer" containerID="cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.563776 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-scheduler-0" podStartSLOduration=4.577712753 podStartE2EDuration="5.563754458s" podCreationTimestamp="2026-03-14 09:54:26 +0000 UTC" firstStartedPulling="2026-03-14 09:54:27.723588303 +0000 UTC m=+2635.036199431" lastFinishedPulling="2026-03-14 09:54:28.709630008 +0000 UTC m=+2636.022241136" observedRunningTime="2026-03-14 09:54:31.562635381 +0000 UTC m=+2638.875246509" watchObservedRunningTime="2026-03-14 09:54:31.563754458 +0000 UTC m=+2638.876365596" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.587591 4843 scope.go:117] "RemoveContainer" containerID="d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917" Mar 14 09:54:31 crc kubenswrapper[4843]: E0314 09:54:31.588062 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917\": container with ID starting with d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917 not found: ID does not exist" containerID="d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.588114 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917"} err="failed to get container status \"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917\": rpc error: code = NotFound desc = could not find container \"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917\": container with ID starting with d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917 not found: ID does not exist" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.588150 4843 scope.go:117] "RemoveContainer" containerID="cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0" Mar 14 09:54:31 crc kubenswrapper[4843]: E0314 09:54:31.588527 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0\": container with ID starting with cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0 not found: ID does not exist" containerID="cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.588559 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0"} err="failed to get container status \"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0\": rpc error: code = NotFound desc = could not find container \"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0\": container with ID starting with cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0 not found: ID does not exist" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.588579 4843 scope.go:117] "RemoveContainer" containerID="d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.588877 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917"} err="failed to get container status \"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917\": rpc error: code = NotFound desc = could not find container \"d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917\": container with ID starting with d7b7e770213cb870f7656f75467586d4f2297565e3e496fa9b63b8da4e75b917 not found: ID does not exist" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.588909 4843 scope.go:117] "RemoveContainer" containerID="cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.589703 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0"} err="failed to get container status \"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0\": rpc error: code = NotFound desc = could not find container \"cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0\": container with ID starting with cc9e44f20b3daff6a9deec11e7e11324a075c0a2b8c8590257e873d585ea52f0 not found: ID does not exist" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629731 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-combined-ca-bundle\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629833 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629857 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data-custom\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629891 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38083617-1609-475c-98dc-f475249194bd-logs\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629916 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-scripts\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629973 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfr77\" (UniqueName: \"kubernetes.io/projected/38083617-1609-475c-98dc-f475249194bd-kube-api-access-zfr77\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.629994 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-cert-memcached-mtls\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.630022 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38083617-1609-475c-98dc-f475249194bd-etc-machine-id\") pod \"38083617-1609-475c-98dc-f475249194bd\" (UID: \"38083617-1609-475c-98dc-f475249194bd\") " Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.630529 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38083617-1609-475c-98dc-f475249194bd-logs" (OuterVolumeSpecName: "logs") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.630567 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38083617-1609-475c-98dc-f475249194bd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.631239 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38083617-1609-475c-98dc-f475249194bd-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.631268 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38083617-1609-475c-98dc-f475249194bd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.634876 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.643312 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38083617-1609-475c-98dc-f475249194bd-kube-api-access-zfr77" (OuterVolumeSpecName: "kube-api-access-zfr77") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "kube-api-access-zfr77". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.654122 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.655557 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-scripts" (OuterVolumeSpecName: "scripts") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.694390 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data" (OuterVolumeSpecName: "config-data") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.732763 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.732802 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.732813 4843 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.732824 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.732836 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfr77\" (UniqueName: \"kubernetes.io/projected/38083617-1609-475c-98dc-f475249194bd-kube-api-access-zfr77\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.734411 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "38083617-1609-475c-98dc-f475249194bd" (UID: "38083617-1609-475c-98dc-f475249194bd"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.834896 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/38083617-1609-475c-98dc-f475249194bd-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.869956 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.880632 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.895953 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:31 crc kubenswrapper[4843]: E0314 09:54:31.896306 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.896321 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api" Mar 14 09:54:31 crc kubenswrapper[4843]: E0314 09:54:31.896341 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api-log" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.896347 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api-log" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.896505 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api-log" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.896552 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="38083617-1609-475c-98dc-f475249194bd" containerName="cinder-api" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.897549 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.900738 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-cinder-internal-svc" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.900922 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-cinder-public-svc" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.901789 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-api-config-data" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.908737 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.925965 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:31 crc kubenswrapper[4843]: I0314 09:54:31.976415 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.038987 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8901d-9037-4d8e-97da-b9d80de56af4-logs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039037 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0b8901d-9037-4d8e-97da-b9d80de56af4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039068 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039085 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039124 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039155 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp2dg\" (UniqueName: \"kubernetes.io/projected/d0b8901d-9037-4d8e-97da-b9d80de56af4-kube-api-access-rp2dg\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039198 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039217 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039234 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-scripts\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.039249 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.140578 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp2dg\" (UniqueName: \"kubernetes.io/projected/d0b8901d-9037-4d8e-97da-b9d80de56af4-kube-api-access-rp2dg\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.141158 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.141702 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.141739 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-scripts\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.141761 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.141808 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8901d-9037-4d8e-97da-b9d80de56af4-logs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.141829 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0b8901d-9037-4d8e-97da-b9d80de56af4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.142329 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8901d-9037-4d8e-97da-b9d80de56af4-logs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.142369 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0b8901d-9037-4d8e-97da-b9d80de56af4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.142596 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.142751 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.142925 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.145387 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.146215 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.146438 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.146841 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-scripts\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.146977 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.147590 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-cert-memcached-mtls\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.150226 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.172242 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp2dg\" (UniqueName: \"kubernetes.io/projected/d0b8901d-9037-4d8e-97da-b9d80de56af4-kube-api-access-rp2dg\") pod \"cinder-api-0\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.226137 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.416486 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:32 crc kubenswrapper[4843]: I0314 09:54:32.723688 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:33 crc kubenswrapper[4843]: I0314 09:54:33.358197 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38083617-1609-475c-98dc-f475249194bd" path="/var/lib/kubelet/pods/38083617-1609-475c-98dc-f475249194bd/volumes" Mar 14 09:54:33 crc kubenswrapper[4843]: I0314 09:54:33.556196 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d0b8901d-9037-4d8e-97da-b9d80de56af4","Type":"ContainerStarted","Data":"3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f"} Mar 14 09:54:33 crc kubenswrapper[4843]: I0314 09:54:33.556243 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d0b8901d-9037-4d8e-97da-b9d80de56af4","Type":"ContainerStarted","Data":"ef6d46219391da4afc972f47f610cd47075166bc96be08bf509e7de8c8a1ab32"} Mar 14 09:54:33 crc kubenswrapper[4843]: I0314 09:54:33.675054 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:34 crc kubenswrapper[4843]: I0314 09:54:34.570212 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d0b8901d-9037-4d8e-97da-b9d80de56af4","Type":"ContainerStarted","Data":"535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e"} Mar 14 09:54:34 crc kubenswrapper[4843]: I0314 09:54:34.570551 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:34 crc kubenswrapper[4843]: I0314 09:54:34.597034 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-api-0" podStartSLOduration=3.597017816 podStartE2EDuration="3.597017816s" podCreationTimestamp="2026-03-14 09:54:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:54:34.594396242 +0000 UTC m=+2641.907007380" watchObservedRunningTime="2026-03-14 09:54:34.597017816 +0000 UTC m=+2641.909628944" Mar 14 09:54:34 crc kubenswrapper[4843]: I0314 09:54:34.964431 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:36 crc kubenswrapper[4843]: I0314 09:54:36.235866 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.165499 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.218250 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.267227 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.336245 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.418035 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.596903 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="cinder-backup" containerID="cri-o://731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d" gracePeriod=30 Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.597002 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="probe" containerID="cri-o://c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5" gracePeriod=30 Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.597089 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="probe" containerID="cri-o://6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f" gracePeriod=30 Mar 14 09:54:37 crc kubenswrapper[4843]: I0314 09:54:37.597112 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="cinder-scheduler" containerID="cri-o://177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8" gracePeriod=30 Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.634856 4843 generic.go:334] "Generic (PLEG): container finished" podID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerID="c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5" exitCode=0 Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.634976 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"d220fd8b-4d05-498c-8a78-7035f12b054c","Type":"ContainerDied","Data":"c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5"} Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.637190 4843 generic.go:334] "Generic (PLEG): container finished" podID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerID="6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f" exitCode=0 Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.637235 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090","Type":"ContainerDied","Data":"6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f"} Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.651974 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.921821 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:38 crc kubenswrapper[4843]: I0314 09:54:38.922337 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="c847b0bd-b26d-4d23-a465-15d57f1a968e" containerName="watcher-decision-engine" containerID="cri-o://6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4" gracePeriod=30 Mar 14 09:54:39 crc kubenswrapper[4843]: I0314 09:54:39.934611 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:39 crc kubenswrapper[4843]: I0314 09:54:39.968755 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:39 crc kubenswrapper[4843]: I0314 09:54:39.969016 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-central-agent" containerID="cri-o://8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b" gracePeriod=30 Mar 14 09:54:39 crc kubenswrapper[4843]: I0314 09:54:39.969094 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-notification-agent" containerID="cri-o://11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822" gracePeriod=30 Mar 14 09:54:39 crc kubenswrapper[4843]: I0314 09:54:39.969078 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="proxy-httpd" containerID="cri-o://f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a" gracePeriod=30 Mar 14 09:54:39 crc kubenswrapper[4843]: I0314 09:54:39.969079 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="sg-core" containerID="cri-o://24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31" gracePeriod=30 Mar 14 09:54:40 crc kubenswrapper[4843]: I0314 09:54:40.654905 4843 generic.go:334] "Generic (PLEG): container finished" podID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerID="f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a" exitCode=0 Mar 14 09:54:40 crc kubenswrapper[4843]: I0314 09:54:40.654955 4843 generic.go:334] "Generic (PLEG): container finished" podID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerID="24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31" exitCode=2 Mar 14 09:54:40 crc kubenswrapper[4843]: I0314 09:54:40.654964 4843 generic.go:334] "Generic (PLEG): container finished" podID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerID="8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b" exitCode=0 Mar 14 09:54:40 crc kubenswrapper[4843]: I0314 09:54:40.654983 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerDied","Data":"f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a"} Mar 14 09:54:40 crc kubenswrapper[4843]: I0314 09:54:40.655027 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerDied","Data":"24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31"} Mar 14 09:54:40 crc kubenswrapper[4843]: I0314 09:54:40.655037 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerDied","Data":"8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b"} Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.122171 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_c847b0bd-b26d-4d23-a465-15d57f1a968e/watcher-decision-engine/0.log" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.501610 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.577350 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.615905 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.615960 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c847b0bd-b26d-4d23-a465-15d57f1a968e-logs\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.616025 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-combined-ca-bundle\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.616130 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-config-data\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.616151 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h8lz\" (UniqueName: \"kubernetes.io/projected/c847b0bd-b26d-4d23-a465-15d57f1a968e-kube-api-access-2h8lz\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.616170 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-custom-prometheus-ca\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.616442 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c847b0bd-b26d-4d23-a465-15d57f1a968e-logs" (OuterVolumeSpecName: "logs") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.616664 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c847b0bd-b26d-4d23-a465-15d57f1a968e-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.622652 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c847b0bd-b26d-4d23-a465-15d57f1a968e-kube-api-access-2h8lz" (OuterVolumeSpecName: "kube-api-access-2h8lz") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "kube-api-access-2h8lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.662608 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.674366 4843 generic.go:334] "Generic (PLEG): container finished" podID="c847b0bd-b26d-4d23-a465-15d57f1a968e" containerID="6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4" exitCode=0 Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.674454 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"c847b0bd-b26d-4d23-a465-15d57f1a968e","Type":"ContainerDied","Data":"6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4"} Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.674483 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"c847b0bd-b26d-4d23-a465-15d57f1a968e","Type":"ContainerDied","Data":"be5e87ab6227d3a6088a6080a272ac1323895c4cadfb7134b0c1a67f14392a18"} Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.674499 4843 scope.go:117] "RemoveContainer" containerID="6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.674511 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.680117 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.682203 4843 generic.go:334] "Generic (PLEG): container finished" podID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerID="11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822" exitCode=0 Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.682240 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerDied","Data":"11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822"} Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.682265 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"6cce67ea-d82a-441b-a52b-eb025169a40b","Type":"ContainerDied","Data":"2fd1df63b051ff30d4e09eaf26e80cdf9d238d6d701b0178eebd11cefe0dc331"} Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.682320 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.687670 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-config-data" (OuterVolumeSpecName: "config-data") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.716923 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717273 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-run-httpd\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717348 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-scripts\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717547 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls\") pod \"c847b0bd-b26d-4d23-a465-15d57f1a968e\" (UID: \"c847b0bd-b26d-4d23-a465-15d57f1a968e\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717569 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-combined-ca-bundle\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717606 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-ceilometer-tls-certs\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717709 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: W0314 09:54:41.717719 4843 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c847b0bd-b26d-4d23-a465-15d57f1a968e/volumes/kubernetes.io~secret/cert-memcached-mtls Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717742 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhc7p\" (UniqueName: \"kubernetes.io/projected/6cce67ea-d82a-441b-a52b-eb025169a40b-kube-api-access-mhc7p\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717750 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "c847b0bd-b26d-4d23-a465-15d57f1a968e" (UID: "c847b0bd-b26d-4d23-a465-15d57f1a968e"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717780 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-log-httpd\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717815 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-sg-core-conf-yaml\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.717830 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-config-data\") pod \"6cce67ea-d82a-441b-a52b-eb025169a40b\" (UID: \"6cce67ea-d82a-441b-a52b-eb025169a40b\") " Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.718184 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.718198 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.718207 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h8lz\" (UniqueName: \"kubernetes.io/projected/c847b0bd-b26d-4d23-a465-15d57f1a968e-kube-api-access-2h8lz\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.718216 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.718223 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.718231 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/c847b0bd-b26d-4d23-a465-15d57f1a968e-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.719582 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.720535 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cce67ea-d82a-441b-a52b-eb025169a40b-kube-api-access-mhc7p" (OuterVolumeSpecName: "kube-api-access-mhc7p") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "kube-api-access-mhc7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.720981 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-scripts" (OuterVolumeSpecName: "scripts") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.741134 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.754467 4843 scope.go:117] "RemoveContainer" containerID="6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4" Mar 14 09:54:41 crc kubenswrapper[4843]: E0314 09:54:41.754829 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4\": container with ID starting with 6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4 not found: ID does not exist" containerID="6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.754901 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4"} err="failed to get container status \"6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4\": rpc error: code = NotFound desc = could not find container \"6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4\": container with ID starting with 6380129e22ac9cebcfe41ede28752789a2143a640f7305be644122c4ae9a95b4 not found: ID does not exist" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.754925 4843 scope.go:117] "RemoveContainer" containerID="f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.760420 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.785830 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.818079 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-config-data" (OuterVolumeSpecName: "config-data") pod "6cce67ea-d82a-441b-a52b-eb025169a40b" (UID: "6cce67ea-d82a-441b-a52b-eb025169a40b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819790 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819814 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819841 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819854 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhc7p\" (UniqueName: \"kubernetes.io/projected/6cce67ea-d82a-441b-a52b-eb025169a40b-kube-api-access-mhc7p\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819863 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6cce67ea-d82a-441b-a52b-eb025169a40b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819871 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.819879 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cce67ea-d82a-441b-a52b-eb025169a40b-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.852018 4843 scope.go:117] "RemoveContainer" containerID="24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.875945 4843 scope.go:117] "RemoveContainer" containerID="11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822" Mar 14 09:54:41 crc kubenswrapper[4843]: I0314 09:54:41.977105 4843 scope.go:117] "RemoveContainer" containerID="8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.080343 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.092492 4843 scope.go:117] "RemoveContainer" containerID="f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.093540 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a\": container with ID starting with f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a not found: ID does not exist" containerID="f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.093574 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a"} err="failed to get container status \"f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a\": rpc error: code = NotFound desc = could not find container \"f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a\": container with ID starting with f50368b9ce9e4d8a6c748de9e84fa7847ed66af5555f5b4ea5665c01c131cf8a not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.093615 4843 scope.go:117] "RemoveContainer" containerID="24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.093915 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31\": container with ID starting with 24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31 not found: ID does not exist" containerID="24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.093931 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31"} err="failed to get container status \"24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31\": rpc error: code = NotFound desc = could not find container \"24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31\": container with ID starting with 24f1249bf735ff705d63418747bf9361e7a0151d7bc45d481b72a5a9fe15bd31 not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.093946 4843 scope.go:117] "RemoveContainer" containerID="11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.094163 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822\": container with ID starting with 11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822 not found: ID does not exist" containerID="11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.094179 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822"} err="failed to get container status \"11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822\": rpc error: code = NotFound desc = could not find container \"11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822\": container with ID starting with 11f489d6813df41adeee43db310f6a33dee261bfb7f5836ab2f67edce73f8822 not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.094208 4843 scope.go:117] "RemoveContainer" containerID="8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.094391 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b\": container with ID starting with 8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b not found: ID does not exist" containerID="8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.094432 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b"} err="failed to get container status \"8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b\": rpc error: code = NotFound desc = could not find container \"8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b\": container with ID starting with 8df9a2facd6a9ddb07940d709a1c67fcbc974cb42271810b7bec2655e5f0a59b not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.101683 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.155346 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.162350 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.166417 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.166782 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="proxy-httpd" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.166794 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="proxy-httpd" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.166808 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-central-agent" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.166813 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-central-agent" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.166826 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c847b0bd-b26d-4d23-a465-15d57f1a968e" containerName="watcher-decision-engine" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.166833 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c847b0bd-b26d-4d23-a465-15d57f1a968e" containerName="watcher-decision-engine" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.166848 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-notification-agent" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.166853 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-notification-agent" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.166872 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="sg-core" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.166878 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="sg-core" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.167012 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-central-agent" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.167022 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="ceilometer-notification-agent" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.167031 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c847b0bd-b26d-4d23-a465-15d57f1a968e" containerName="watcher-decision-engine" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.167038 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="proxy-httpd" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.167048 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" containerName="sg-core" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.167600 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.172650 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.211555 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.240336 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.243935 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.249776 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.250020 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.257643 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.258016 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.262094 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335077 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335366 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335454 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pq2k\" (UniqueName: \"kubernetes.io/projected/2db0e8e9-cb24-4896-bc66-31cc71d553cd-kube-api-access-7pq2k\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335523 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335549 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-scripts\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335596 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ef26b5-077f-4419-87a4-b1100c18f189-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335692 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-log-httpd\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335720 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335779 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxlsq\" (UniqueName: \"kubernetes.io/projected/b1ef26b5-077f-4419-87a4-b1100c18f189-kube-api-access-kxlsq\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335814 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335867 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335908 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-config-data\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335951 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.335978 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-run-httpd\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.437016 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.437130 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-cert-memcached-mtls\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.437157 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-scripts\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.437179 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data-custom\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.438176 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-combined-ca-bundle\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.438563 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-etc-machine-id\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.438634 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snrbq\" (UniqueName: \"kubernetes.io/projected/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-kube-api-access-snrbq\") pod \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\" (UID: \"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.438702 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.438917 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.438965 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-run-httpd\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439023 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439054 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439088 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pq2k\" (UniqueName: \"kubernetes.io/projected/2db0e8e9-cb24-4896-bc66-31cc71d553cd-kube-api-access-7pq2k\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439138 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439165 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-scripts\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439201 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ef26b5-077f-4419-87a4-b1100c18f189-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439343 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-log-httpd\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439381 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439429 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxlsq\" (UniqueName: \"kubernetes.io/projected/b1ef26b5-077f-4419-87a4-b1100c18f189-kube-api-access-kxlsq\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439474 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439517 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439560 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-config-data\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439629 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.439685 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-run-httpd\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.441175 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ef26b5-077f-4419-87a4-b1100c18f189-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.441304 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-scripts" (OuterVolumeSpecName: "scripts") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.441539 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-log-httpd\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.451298 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.451351 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-scripts\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.451455 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.452169 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.453747 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.460324 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-config-data\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.462432 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-kube-api-access-snrbq" (OuterVolumeSpecName: "kube-api-access-snrbq") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "kube-api-access-snrbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.462882 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.473722 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxlsq\" (UniqueName: \"kubernetes.io/projected/b1ef26b5-077f-4419-87a4-b1100c18f189-kube-api-access-kxlsq\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.473799 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.473826 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pq2k\" (UniqueName: \"kubernetes.io/projected/2db0e8e9-cb24-4896-bc66-31cc71d553cd-kube-api-access-7pq2k\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.476710 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.484293 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.519326 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.529376 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.547095 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.547132 4843 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.547143 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.547152 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snrbq\" (UniqueName: \"kubernetes.io/projected/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-kube-api-access-snrbq\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.571106 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data" (OuterVolumeSpecName: "config-data") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.605085 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.615477 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.641115 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" (UID: "d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647633 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-brick\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647683 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-machine-id\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647737 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-run\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647761 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647815 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-lib-modules\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647801 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647845 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-scripts\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647871 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-sys\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647891 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-cinder\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.647911 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-cert-memcached-mtls\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648167 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-sys" (OuterVolumeSpecName: "sys") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648228 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-run" (OuterVolumeSpecName: "run") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648251 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648311 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648518 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-iscsi\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648620 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data-custom\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648639 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648687 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648720 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-dev\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648741 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-combined-ca-bundle\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648756 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-nvme\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648782 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-lib-cinder\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.648801 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdvft\" (UniqueName: \"kubernetes.io/projected/d220fd8b-4d05-498c-8a78-7035f12b054c-kube-api-access-cdvft\") pod \"d220fd8b-4d05-498c-8a78-7035f12b054c\" (UID: \"d220fd8b-4d05-498c-8a78-7035f12b054c\") " Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649535 4843 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649554 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649563 4843 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-run\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649573 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649570 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649581 4843 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-lib-modules\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649591 4843 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-sys\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649599 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649605 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-dev" (OuterVolumeSpecName: "dev") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649606 4843 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649639 4843 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.649662 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.652518 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-scripts" (OuterVolumeSpecName: "scripts") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.652863 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d220fd8b-4d05-498c-8a78-7035f12b054c-kube-api-access-cdvft" (OuterVolumeSpecName: "kube-api-access-cdvft") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "kube-api-access-cdvft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.654398 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.747004 4843 generic.go:334] "Generic (PLEG): container finished" podID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerID="731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d" exitCode=0 Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.747311 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"d220fd8b-4d05-498c-8a78-7035f12b054c","Type":"ContainerDied","Data":"731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d"} Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.747340 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"d220fd8b-4d05-498c-8a78-7035f12b054c","Type":"ContainerDied","Data":"8ae46251efa386aac6be482e0e5083b623fab593e69cb74181a435588374faf0"} Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.747356 4843 scope.go:117] "RemoveContainer" containerID="c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.747524 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.751149 4843 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.751171 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdvft\" (UniqueName: \"kubernetes.io/projected/d220fd8b-4d05-498c-8a78-7035f12b054c-kube-api-access-cdvft\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.751181 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.751190 4843 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.751201 4843 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-dev\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.751209 4843 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d220fd8b-4d05-498c-8a78-7035f12b054c-etc-nvme\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.762326 4843 generic.go:334] "Generic (PLEG): container finished" podID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerID="177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8" exitCode=0 Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.762359 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090","Type":"ContainerDied","Data":"177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8"} Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.762384 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090","Type":"ContainerDied","Data":"7e68db2d8324f6939668ec1c9911a9f9aacfc34b93d67213f9defb4b09eba8cd"} Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.762448 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.770781 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.792551 4843 scope.go:117] "RemoveContainer" containerID="731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.800819 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.807454 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.839201 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.841381 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="cinder-backup" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841403 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="cinder-backup" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.841420 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="cinder-scheduler" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841426 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="cinder-scheduler" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.841434 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="probe" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841441 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="probe" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.841467 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="probe" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841473 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="probe" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841614 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="probe" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841626 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="cinder-backup" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841640 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" containerName="cinder-scheduler" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.841650 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" containerName="probe" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.842527 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.846807 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-scheduler-config-data" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.846970 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.855860 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.866473 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.876021 4843 scope.go:117] "RemoveContainer" containerID="c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.877150 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5\": container with ID starting with c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5 not found: ID does not exist" containerID="c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.877178 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5"} err="failed to get container status \"c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5\": rpc error: code = NotFound desc = could not find container \"c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5\": container with ID starting with c8aba60d1e12d0958be296fb4cc4dfc70324ef0942385f3873e90ab699a808f5 not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.877199 4843 scope.go:117] "RemoveContainer" containerID="731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.877634 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d\": container with ID starting with 731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d not found: ID does not exist" containerID="731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.877659 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d"} err="failed to get container status \"731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d\": rpc error: code = NotFound desc = could not find container \"731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d\": container with ID starting with 731964b6d91ca8fe9dc32dabb4874bfdb4acf073a568155ee3ac0f948f899b0d not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.877674 4843 scope.go:117] "RemoveContainer" containerID="6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.889506 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data" (OuterVolumeSpecName: "config-data") pod "d220fd8b-4d05-498c-8a78-7035f12b054c" (UID: "d220fd8b-4d05-498c-8a78-7035f12b054c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.914603 4843 scope.go:117] "RemoveContainer" containerID="177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.954177 4843 scope.go:117] "RemoveContainer" containerID="6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.956499 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f\": container with ID starting with 6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f not found: ID does not exist" containerID="6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956530 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f"} err="failed to get container status \"6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f\": rpc error: code = NotFound desc = could not find container \"6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f\": container with ID starting with 6de82e3965286f2b243a4414057a64d15a6985531b28114c4fc396b65e86e36f not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956550 4843 scope.go:117] "RemoveContainer" containerID="177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956801 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956855 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: E0314 09:54:42.956854 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8\": container with ID starting with 177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8 not found: ID does not exist" containerID="177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956891 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8"} err="failed to get container status \"177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8\": rpc error: code = NotFound desc = could not find container \"177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8\": container with ID starting with 177c92302956235f50f99002a89df38c6fbe540b89bb8c1982e6f0da234d8af8 not found: ID does not exist" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956918 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-scripts\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956938 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l42rc\" (UniqueName: \"kubernetes.io/projected/7ab00353-e46e-455f-98df-a89f1192edc0-kube-api-access-l42rc\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.956983 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.957016 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.957054 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ab00353-e46e-455f-98df-a89f1192edc0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.957110 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.957123 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d220fd8b-4d05-498c-8a78-7035f12b054c-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:42 crc kubenswrapper[4843]: I0314 09:54:42.964610 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:42 crc kubenswrapper[4843]: W0314 09:54:42.965386 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1ef26b5_077f_4419_87a4_b1100c18f189.slice/crio-df975bdda7da46be44e7b401c0eab148480a374817a425b05befc9bef43d1cad WatchSource:0}: Error finding container df975bdda7da46be44e7b401c0eab148480a374817a425b05befc9bef43d1cad: Status 404 returned error can't find the container with id df975bdda7da46be44e7b401c0eab148480a374817a425b05befc9bef43d1cad Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061171 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061231 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061296 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ab00353-e46e-455f-98df-a89f1192edc0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061332 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061355 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061381 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-scripts\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.061396 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l42rc\" (UniqueName: \"kubernetes.io/projected/7ab00353-e46e-455f-98df-a89f1192edc0-kube-api-access-l42rc\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.063786 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ab00353-e46e-455f-98df-a89f1192edc0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.066868 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-scripts\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.069071 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.069169 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.069263 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.077973 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-cert-memcached-mtls\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.083108 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l42rc\" (UniqueName: \"kubernetes.io/projected/7ab00353-e46e-455f-98df-a89f1192edc0-kube-api-access-l42rc\") pod \"cinder-scheduler-0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.109360 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.127592 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.146355 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.148003 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.150512 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cinder-backup-config-data" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.158153 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.183320 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.236263 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266069 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266105 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266145 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-lib-modules\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266166 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-sys\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266184 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266259 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-run\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266299 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266319 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266335 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-dev\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266356 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266371 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266394 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266416 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbslx\" (UniqueName: \"kubernetes.io/projected/902898bd-0b20-4fd0-991a-58c64f9b3aa0-kube-api-access-wbslx\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266445 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-scripts\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266460 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.266486 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370224 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-run\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370605 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370630 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370649 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-dev\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370678 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370701 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370730 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370753 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbslx\" (UniqueName: \"kubernetes.io/projected/902898bd-0b20-4fd0-991a-58c64f9b3aa0-kube-api-access-wbslx\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370788 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-scripts\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370805 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370833 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370862 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370877 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370908 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-lib-modules\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370933 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-sys\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.370952 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.371081 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.371124 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-run\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.371161 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.371197 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.371224 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-dev\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.371428 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.373038 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.373084 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.374683 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cce67ea-d82a-441b-a52b-eb025169a40b" path="/var/lib/kubelet/pods/6cce67ea-d82a-441b-a52b-eb025169a40b/volumes" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.375361 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-lib-modules\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.375433 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-sys\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.379975 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.381113 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.381169 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.382143 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-scripts\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.387468 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c847b0bd-b26d-4d23-a465-15d57f1a968e" path="/var/lib/kubelet/pods/c847b0bd-b26d-4d23-a465-15d57f1a968e/volumes" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.388236 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d220fd8b-4d05-498c-8a78-7035f12b054c" path="/var/lib/kubelet/pods/d220fd8b-4d05-498c-8a78-7035f12b054c/volumes" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.388793 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090" path="/var/lib/kubelet/pods/d3f5c75d-5a6c-4a80-8ad4-e5046b5e9090/volumes" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.396650 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbslx\" (UniqueName: \"kubernetes.io/projected/902898bd-0b20-4fd0-991a-58c64f9b3aa0-kube-api-access-wbslx\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.410806 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-cert-memcached-mtls\") pod \"cinder-backup-0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.619042 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.778738 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.798633 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerStarted","Data":"806bef1ae71e1ee00fb511f39d633290e1253c3e1c3bd393eec01be75d412809"} Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.802170 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"b1ef26b5-077f-4419-87a4-b1100c18f189","Type":"ContainerStarted","Data":"80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862"} Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.802204 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"b1ef26b5-077f-4419-87a4-b1100c18f189","Type":"ContainerStarted","Data":"df975bdda7da46be44e7b401c0eab148480a374817a425b05befc9bef43d1cad"} Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.830476 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=1.830403169 podStartE2EDuration="1.830403169s" podCreationTimestamp="2026-03-14 09:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:54:43.828301218 +0000 UTC m=+2651.140912346" watchObservedRunningTime="2026-03-14 09:54:43.830403169 +0000 UTC m=+2651.143014297" Mar 14 09:54:43 crc kubenswrapper[4843]: I0314 09:54:43.943264 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.692196 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.785963 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.825300 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"902898bd-0b20-4fd0-991a-58c64f9b3aa0","Type":"ContainerStarted","Data":"ab0028aba79a6aec38553cded7be802844a02b5718686853475b81800f0a51ca"} Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.825341 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"902898bd-0b20-4fd0-991a-58c64f9b3aa0","Type":"ContainerStarted","Data":"80889a134df7c58d16daa50cb7cb09b9f3cb65dc4c22b0ed1a95e8238fe721bf"} Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.825351 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"902898bd-0b20-4fd0-991a-58c64f9b3aa0","Type":"ContainerStarted","Data":"91d4577ddca889acb79414f7ca56ab144730c671138ece375381dccbc239e6fe"} Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.827109 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerStarted","Data":"3343f6f7dbf0e742dee958b6e6276c71a7074d73a125ecec3c9a21c4f3ef71d4"} Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.860072 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"7ab00353-e46e-455f-98df-a89f1192edc0","Type":"ContainerStarted","Data":"8f8524a861ca84f25b68de2f13eebedf2cdef3501e1555f9efd60bc6a2c1c648"} Mar 14 09:54:44 crc kubenswrapper[4843]: I0314 09:54:44.860313 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"7ab00353-e46e-455f-98df-a89f1192edc0","Type":"ContainerStarted","Data":"34cfd5f13d36cf182de93a46ae3ea11b7ef447417521ac96076cc046c848dd7b"} Mar 14 09:54:45 crc kubenswrapper[4843]: I0314 09:54:45.871665 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerStarted","Data":"edfe00d1ad21ae85435ae3aae021d9a8d5f52231f61849e008f63c0d1ad34db6"} Mar 14 09:54:45 crc kubenswrapper[4843]: I0314 09:54:45.875264 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"7ab00353-e46e-455f-98df-a89f1192edc0","Type":"ContainerStarted","Data":"f403133beb54c29f3eaf7d90bd4982ce8a220d34a36d2861d39cb44edd63d32c"} Mar 14 09:54:45 crc kubenswrapper[4843]: I0314 09:54:45.897809 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-scheduler-0" podStartSLOduration=3.897790425 podStartE2EDuration="3.897790425s" podCreationTimestamp="2026-03-14 09:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:54:45.896724379 +0000 UTC m=+2653.209335507" watchObservedRunningTime="2026-03-14 09:54:45.897790425 +0000 UTC m=+2653.210401553" Mar 14 09:54:45 crc kubenswrapper[4843]: I0314 09:54:45.903764 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/cinder-backup-0" podStartSLOduration=2.90374492 podStartE2EDuration="2.90374492s" podCreationTimestamp="2026-03-14 09:54:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:54:44.852439586 +0000 UTC m=+2652.165050714" watchObservedRunningTime="2026-03-14 09:54:45.90374492 +0000 UTC m=+2653.216356048" Mar 14 09:54:45 crc kubenswrapper[4843]: I0314 09:54:45.971883 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:46 crc kubenswrapper[4843]: I0314 09:54:46.885253 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerStarted","Data":"874949fbd1e7ec53ac8466b5b42f25cb17de84f553f45b3fd652d590b732b24d"} Mar 14 09:54:47 crc kubenswrapper[4843]: I0314 09:54:47.149030 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:47 crc kubenswrapper[4843]: I0314 09:54:47.895580 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerStarted","Data":"033dd4c94971b19eadcfede69485c051ebc6726f2fca9e50c4a52cd3ed7f139d"} Mar 14 09:54:47 crc kubenswrapper[4843]: I0314 09:54:47.895942 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:47 crc kubenswrapper[4843]: I0314 09:54:47.926945 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.827487637 podStartE2EDuration="5.926929283s" podCreationTimestamp="2026-03-14 09:54:42 +0000 UTC" firstStartedPulling="2026-03-14 09:54:43.255784218 +0000 UTC m=+2650.568395336" lastFinishedPulling="2026-03-14 09:54:47.355225854 +0000 UTC m=+2654.667836982" observedRunningTime="2026-03-14 09:54:47.920577377 +0000 UTC m=+2655.233188505" watchObservedRunningTime="2026-03-14 09:54:47.926929283 +0000 UTC m=+2655.239540411" Mar 14 09:54:48 crc kubenswrapper[4843]: I0314 09:54:48.184382 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:48 crc kubenswrapper[4843]: I0314 09:54:48.343602 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:48 crc kubenswrapper[4843]: I0314 09:54:48.620610 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:49 crc kubenswrapper[4843]: I0314 09:54:49.596554 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:50 crc kubenswrapper[4843]: I0314 09:54:50.791022 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.165059 4843 scope.go:117] "RemoveContainer" containerID="4559a397bebfabe03c183b58325b36823c289dd3bad2c87f701bf5905c15bb2d" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.226478 4843 scope.go:117] "RemoveContainer" containerID="6eda1fb53b7eff4793ea91205dc643197aebfb1a3e38508b7edc79e87efa5cd0" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.281933 4843 scope.go:117] "RemoveContainer" containerID="d4693b83f32a63fb2df86bdd6281bbde1cf2f96696fdc87f6ee5addc316288ed" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.311247 4843 scope.go:117] "RemoveContainer" containerID="086156d7f2f3de1323812bef986b946e34740a54f026f70c9aec24f3506df087" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.343827 4843 scope.go:117] "RemoveContainer" containerID="6efbdf372a841f2ec509a824c2acd7bb46cdacdc022eab68f0bb7ec7dc9c8b23" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.370677 4843 scope.go:117] "RemoveContainer" containerID="58aca775f2e0644db2b9fc28254a813d0792d753edb7ace172581e32735e9b01" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.402212 4843 scope.go:117] "RemoveContainer" containerID="3295c4030a7b85687715d84059a654373b7cf06c816f87b2acb9ef54d96266a6" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.434647 4843 scope.go:117] "RemoveContainer" containerID="da4b54856f14ad3b8dd74f9c2d512a81dbf3786bd07508c800043ded8f64892f" Mar 14 09:54:51 crc kubenswrapper[4843]: I0314 09:54:51.472490 4843 scope.go:117] "RemoveContainer" containerID="81815b8418cdf4f5b7d4d8f9dc00bc6db437d8a38581b33df5df65168f303900" Mar 14 09:54:52 crc kubenswrapper[4843]: I0314 09:54:52.067214 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:52 crc kubenswrapper[4843]: I0314 09:54:52.606186 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:52 crc kubenswrapper[4843]: I0314 09:54:52.634716 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:52 crc kubenswrapper[4843]: I0314 09:54:52.944770 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:52 crc kubenswrapper[4843]: I0314 09:54:52.985336 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:54:53 crc kubenswrapper[4843]: I0314 09:54:53.348349 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:53 crc kubenswrapper[4843]: I0314 09:54:53.391721 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:53 crc kubenswrapper[4843]: I0314 09:54:53.815477 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:54 crc kubenswrapper[4843]: I0314 09:54:54.554253 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:54 crc kubenswrapper[4843]: I0314 09:54:54.795187 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:54 crc kubenswrapper[4843]: I0314 09:54:54.976361 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-9kdvr"] Mar 14 09:54:54 crc kubenswrapper[4843]: I0314 09:54:54.982897 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-db-sync-9kdvr"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.010001 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.010231 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="cinder-scheduler" containerID="cri-o://8f8524a861ca84f25b68de2f13eebedf2cdef3501e1555f9efd60bc6a2c1c648" gracePeriod=30 Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.010330 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-scheduler-0" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="probe" containerID="cri-o://f403133beb54c29f3eaf7d90bd4982ce8a220d34a36d2861d39cb44edd63d32c" gracePeriod=30 Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.031592 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.031832 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="cinder-backup" containerID="cri-o://80889a134df7c58d16daa50cb7cb09b9f3cb65dc4c22b0ed1a95e8238fe721bf" gracePeriod=30 Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.031923 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-backup-0" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="probe" containerID="cri-o://ab0028aba79a6aec38553cded7be802844a02b5718686853475b81800f0a51ca" gracePeriod=30 Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.081942 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.082237 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api-log" containerID="cri-o://3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f" gracePeriod=30 Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.083905 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/cinder-api-0" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api" containerID="cri-o://535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e" gracePeriod=30 Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.098975 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/cinder9191-account-delete-4xsz9"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.100256 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.112438 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder9191-account-delete-4xsz9"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.181218 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl6f2\" (UniqueName: \"kubernetes.io/projected/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-kube-api-access-xl6f2\") pod \"cinder9191-account-delete-4xsz9\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.181348 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-operator-scripts\") pod \"cinder9191-account-delete-4xsz9\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.282866 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl6f2\" (UniqueName: \"kubernetes.io/projected/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-kube-api-access-xl6f2\") pod \"cinder9191-account-delete-4xsz9\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.282993 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-operator-scripts\") pod \"cinder9191-account-delete-4xsz9\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.283811 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-operator-scripts\") pod \"cinder9191-account-delete-4xsz9\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.306924 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl6f2\" (UniqueName: \"kubernetes.io/projected/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-kube-api-access-xl6f2\") pod \"cinder9191-account-delete-4xsz9\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.348053 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6" path="/var/lib/kubelet/pods/acaf6bc4-ecac-43b6-83c3-3509cf5a3fe6/volumes" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.434895 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.911836 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/cinder9191-account-delete-4xsz9"] Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.981374 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:55 crc kubenswrapper[4843]: I0314 09:54:55.996191 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" event={"ID":"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde","Type":"ContainerStarted","Data":"fec16bbe7a02802f21b87ff0cb0942827dd31b3af887fc58909fa2a522e0b10c"} Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.006032 4843 generic.go:334] "Generic (PLEG): container finished" podID="7ab00353-e46e-455f-98df-a89f1192edc0" containerID="f403133beb54c29f3eaf7d90bd4982ce8a220d34a36d2861d39cb44edd63d32c" exitCode=0 Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.006099 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"7ab00353-e46e-455f-98df-a89f1192edc0","Type":"ContainerDied","Data":"f403133beb54c29f3eaf7d90bd4982ce8a220d34a36d2861d39cb44edd63d32c"} Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.013261 4843 generic.go:334] "Generic (PLEG): container finished" podID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerID="ab0028aba79a6aec38553cded7be802844a02b5718686853475b81800f0a51ca" exitCode=0 Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.013333 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"902898bd-0b20-4fd0-991a-58c64f9b3aa0","Type":"ContainerDied","Data":"ab0028aba79a6aec38553cded7be802844a02b5718686853475b81800f0a51ca"} Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.021378 4843 generic.go:334] "Generic (PLEG): container finished" podID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerID="3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f" exitCode=143 Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.021424 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d0b8901d-9037-4d8e-97da-b9d80de56af4","Type":"ContainerDied","Data":"3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f"} Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.773074 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:54:56 crc kubenswrapper[4843]: I0314 09:54:56.773882 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="b1ef26b5-077f-4419-87a4-b1100c18f189" containerName="watcher-decision-engine" containerID="cri-o://80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862" gracePeriod=30 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.031653 4843 generic.go:334] "Generic (PLEG): container finished" podID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerID="80889a134df7c58d16daa50cb7cb09b9f3cb65dc4c22b0ed1a95e8238fe721bf" exitCode=0 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.031750 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"902898bd-0b20-4fd0-991a-58c64f9b3aa0","Type":"ContainerDied","Data":"80889a134df7c58d16daa50cb7cb09b9f3cb65dc4c22b0ed1a95e8238fe721bf"} Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.033500 4843 generic.go:334] "Generic (PLEG): container finished" podID="f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" containerID="47018f9f60d648e7cb8a81aa957e48673fee220b341a002cbd5db188bea1b9e1" exitCode=0 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.033556 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" event={"ID":"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde","Type":"ContainerDied","Data":"47018f9f60d648e7cb8a81aa957e48673fee220b341a002cbd5db188bea1b9e1"} Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.036064 4843 generic.go:334] "Generic (PLEG): container finished" podID="7ab00353-e46e-455f-98df-a89f1192edc0" containerID="8f8524a861ca84f25b68de2f13eebedf2cdef3501e1555f9efd60bc6a2c1c648" exitCode=0 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.036101 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"7ab00353-e46e-455f-98df-a89f1192edc0","Type":"ContainerDied","Data":"8f8524a861ca84f25b68de2f13eebedf2cdef3501e1555f9efd60bc6a2c1c648"} Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.036123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-scheduler-0" event={"ID":"7ab00353-e46e-455f-98df-a89f1192edc0","Type":"ContainerDied","Data":"34cfd5f13d36cf182de93a46ae3ea11b7ef447417521ac96076cc046c848dd7b"} Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.036136 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34cfd5f13d36cf182de93a46ae3ea11b7ef447417521ac96076cc046c848dd7b" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.062946 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.069435 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115139 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115190 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbslx\" (UniqueName: \"kubernetes.io/projected/902898bd-0b20-4fd0-991a-58c64f9b3aa0-kube-api-access-wbslx\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115221 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-lib-cinder\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115314 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-combined-ca-bundle\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115357 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-cert-memcached-mtls\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115389 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data-custom\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115418 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l42rc\" (UniqueName: \"kubernetes.io/projected/7ab00353-e46e-455f-98df-a89f1192edc0-kube-api-access-l42rc\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115441 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-sys\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115469 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-cert-memcached-mtls\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115507 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-scripts\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115539 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-cinder\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115570 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-brick\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115591 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115618 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-machine-id\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115639 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-run\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115657 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-nvme\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115682 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-iscsi\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115704 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ab00353-e46e-455f-98df-a89f1192edc0-etc-machine-id\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115746 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-dev\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115782 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data-custom\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115822 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-lib-modules\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115847 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-combined-ca-bundle\") pod \"7ab00353-e46e-455f-98df-a89f1192edc0\" (UID: \"7ab00353-e46e-455f-98df-a89f1192edc0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115873 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-scripts\") pod \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\" (UID: \"902898bd-0b20-4fd0-991a-58c64f9b3aa0\") " Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115938 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115938 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-dev" (OuterVolumeSpecName: "dev") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.115999 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ab00353-e46e-455f-98df-a89f1192edc0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116128 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-run" (OuterVolumeSpecName: "run") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116152 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116365 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116035 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116402 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116462 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116755 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.116767 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-sys" (OuterVolumeSpecName: "sys") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117639 4843 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-brick\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117655 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117666 4843 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-run\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117676 4843 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-nvme\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117685 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ab00353-e46e-455f-98df-a89f1192edc0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117694 4843 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-dev\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.117702 4843 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-lib-modules\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.123423 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-scripts" (OuterVolumeSpecName: "scripts") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.135198 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.135386 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902898bd-0b20-4fd0-991a-58c64f9b3aa0-kube-api-access-wbslx" (OuterVolumeSpecName: "kube-api-access-wbslx") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "kube-api-access-wbslx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.136532 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-scripts" (OuterVolumeSpecName: "scripts") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.138872 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.138923 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ab00353-e46e-455f-98df-a89f1192edc0-kube-api-access-l42rc" (OuterVolumeSpecName: "kube-api-access-l42rc") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "kube-api-access-l42rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.180808 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.193228 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.214928 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data" (OuterVolumeSpecName: "config-data") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220213 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220246 4843 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220258 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220303 4843 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-etc-iscsi\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220334 4843 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220392 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220403 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220414 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbslx\" (UniqueName: \"kubernetes.io/projected/902898bd-0b20-4fd0-991a-58c64f9b3aa0-kube-api-access-wbslx\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220424 4843 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220434 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220445 4843 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220456 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l42rc\" (UniqueName: \"kubernetes.io/projected/7ab00353-e46e-455f-98df-a89f1192edc0-kube-api-access-l42rc\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.220466 4843 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/902898bd-0b20-4fd0-991a-58c64f9b3aa0-sys\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.233299 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data" (OuterVolumeSpecName: "config-data") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.270799 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.278739 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "902898bd-0b20-4fd0-991a-58c64f9b3aa0" (UID: "902898bd-0b20-4fd0-991a-58c64f9b3aa0"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.287824 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "7ab00353-e46e-455f-98df-a89f1192edc0" (UID: "7ab00353-e46e-455f-98df-a89f1192edc0"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.322402 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.322434 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/902898bd-0b20-4fd0-991a-58c64f9b3aa0-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.322445 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/7ab00353-e46e-455f-98df-a89f1192edc0-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.414386 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.414654 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-central-agent" containerID="cri-o://3343f6f7dbf0e742dee958b6e6276c71a7074d73a125ecec3c9a21c4f3ef71d4" gracePeriod=30 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.414765 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-notification-agent" containerID="cri-o://edfe00d1ad21ae85435ae3aae021d9a8d5f52231f61849e008f63c0d1ad34db6" gracePeriod=30 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.414805 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="sg-core" containerID="cri-o://874949fbd1e7ec53ac8466b5b42f25cb17de84f553f45b3fd652d590b732b24d" gracePeriod=30 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.414977 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="proxy-httpd" containerID="cri-o://033dd4c94971b19eadcfede69485c051ebc6726f2fca9e50c4a52cd3ed7f139d" gracePeriod=30 Mar 14 09:54:57 crc kubenswrapper[4843]: I0314 09:54:57.424710 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.254:3000/\": EOF" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.047899 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-backup-0" event={"ID":"902898bd-0b20-4fd0-991a-58c64f9b3aa0","Type":"ContainerDied","Data":"91d4577ddca889acb79414f7ca56ab144730c671138ece375381dccbc239e6fe"} Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.049038 4843 scope.go:117] "RemoveContainer" containerID="ab0028aba79a6aec38553cded7be802844a02b5718686853475b81800f0a51ca" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.048177 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-backup-0" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052840 4843 generic.go:334] "Generic (PLEG): container finished" podID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerID="033dd4c94971b19eadcfede69485c051ebc6726f2fca9e50c4a52cd3ed7f139d" exitCode=0 Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052875 4843 generic.go:334] "Generic (PLEG): container finished" podID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerID="874949fbd1e7ec53ac8466b5b42f25cb17de84f553f45b3fd652d590b732b24d" exitCode=2 Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052887 4843 generic.go:334] "Generic (PLEG): container finished" podID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerID="edfe00d1ad21ae85435ae3aae021d9a8d5f52231f61849e008f63c0d1ad34db6" exitCode=0 Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052896 4843 generic.go:334] "Generic (PLEG): container finished" podID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerID="3343f6f7dbf0e742dee958b6e6276c71a7074d73a125ecec3c9a21c4f3ef71d4" exitCode=0 Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052899 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerDied","Data":"033dd4c94971b19eadcfede69485c051ebc6726f2fca9e50c4a52cd3ed7f139d"} Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052937 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerDied","Data":"874949fbd1e7ec53ac8466b5b42f25cb17de84f553f45b3fd652d590b732b24d"} Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052951 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerDied","Data":"edfe00d1ad21ae85435ae3aae021d9a8d5f52231f61849e008f63c0d1ad34db6"} Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.052963 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerDied","Data":"3343f6f7dbf0e742dee958b6e6276c71a7074d73a125ecec3c9a21c4f3ef71d4"} Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.053147 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-scheduler-0" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.086784 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.095062 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-backup-0"] Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.095933 4843 scope.go:117] "RemoveContainer" containerID="80889a134df7c58d16daa50cb7cb09b9f3cb65dc4c22b0ed1a95e8238fe721bf" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.108551 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.122369 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-scheduler-0"] Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.144086 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.249819 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-scripts\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.249942 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pq2k\" (UniqueName: \"kubernetes.io/projected/2db0e8e9-cb24-4896-bc66-31cc71d553cd-kube-api-access-7pq2k\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.250004 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-ceilometer-tls-certs\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.250038 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-config-data\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.250072 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-run-httpd\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.250113 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-combined-ca-bundle\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.250210 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-log-httpd\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.250237 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-sg-core-conf-yaml\") pod \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\" (UID: \"2db0e8e9-cb24-4896-bc66-31cc71d553cd\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.258768 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.260003 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.290508 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db0e8e9-cb24-4896-bc66-31cc71d553cd-kube-api-access-7pq2k" (OuterVolumeSpecName: "kube-api-access-7pq2k") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "kube-api-access-7pq2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.290615 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-scripts" (OuterVolumeSpecName: "scripts") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.341408 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.351853 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.351887 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.351897 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.351906 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pq2k\" (UniqueName: \"kubernetes.io/projected/2db0e8e9-cb24-4896-bc66-31cc71d553cd-kube-api-access-7pq2k\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.351914 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2db0e8e9-cb24-4896-bc66-31cc71d553cd-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.353423 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.448414 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-config-data" (OuterVolumeSpecName: "config-data") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.468360 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.468401 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.509543 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.513986 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.521244 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/cinder-api-0" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.252:8776/healthcheck\": read tcp 10.217.0.2:48760->10.217.0.252:8776: read: connection reset by peer" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.525061 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2db0e8e9-cb24-4896-bc66-31cc71d553cd" (UID: "2db0e8e9-cb24-4896-bc66-31cc71d553cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.569881 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db0e8e9-cb24-4896-bc66-31cc71d553cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.671006 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-operator-scripts\") pod \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.671167 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl6f2\" (UniqueName: \"kubernetes.io/projected/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-kube-api-access-xl6f2\") pod \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\" (UID: \"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde\") " Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.671632 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" (UID: "f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.674178 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-kube-api-access-xl6f2" (OuterVolumeSpecName: "kube-api-access-xl6f2") pod "f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" (UID: "f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde"). InnerVolumeSpecName "kube-api-access-xl6f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.773257 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl6f2\" (UniqueName: \"kubernetes.io/projected/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-kube-api-access-xl6f2\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.773346 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:58 crc kubenswrapper[4843]: I0314 09:54:58.914148 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.450079 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" path="/var/lib/kubelet/pods/7ab00353-e46e-455f-98df-a89f1192edc0/volumes" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.450848 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" path="/var/lib/kubelet/pods/902898bd-0b20-4fd0-991a-58c64f9b3aa0/volumes" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.467897 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.479881 4843 generic.go:334] "Generic (PLEG): container finished" podID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerID="535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e" exitCode=0 Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.480397 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder-api-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.501960 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.517166 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"2db0e8e9-cb24-4896-bc66-31cc71d553cd","Type":"ContainerDied","Data":"806bef1ae71e1ee00fb511f39d633290e1253c3e1c3bd393eec01be75d412809"} Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.517432 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d0b8901d-9037-4d8e-97da-b9d80de56af4","Type":"ContainerDied","Data":"535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e"} Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.517567 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder-api-0" event={"ID":"d0b8901d-9037-4d8e-97da-b9d80de56af4","Type":"ContainerDied","Data":"ef6d46219391da4afc972f47f610cd47075166bc96be08bf509e7de8c8a1ab32"} Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.517686 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/cinder9191-account-delete-4xsz9" event={"ID":"f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde","Type":"ContainerDied","Data":"fec16bbe7a02802f21b87ff0cb0942827dd31b3af887fc58909fa2a522e0b10c"} Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.517757 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fec16bbe7a02802f21b87ff0cb0942827dd31b3af887fc58909fa2a522e0b10c" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.517537 4843 scope.go:117] "RemoveContainer" containerID="033dd4c94971b19eadcfede69485c051ebc6726f2fca9e50c4a52cd3ed7f139d" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531671 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-public-tls-certs\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531747 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-combined-ca-bundle\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531768 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-cert-memcached-mtls\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531813 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-internal-tls-certs\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531833 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data-custom\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531889 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0b8901d-9037-4d8e-97da-b9d80de56af4-etc-machine-id\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.531954 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp2dg\" (UniqueName: \"kubernetes.io/projected/d0b8901d-9037-4d8e-97da-b9d80de56af4-kube-api-access-rp2dg\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.532006 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8901d-9037-4d8e-97da-b9d80de56af4-logs\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.532024 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-scripts\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.532052 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data\") pod \"d0b8901d-9037-4d8e-97da-b9d80de56af4\" (UID: \"d0b8901d-9037-4d8e-97da-b9d80de56af4\") " Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.536939 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0b8901d-9037-4d8e-97da-b9d80de56af4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.537639 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0b8901d-9037-4d8e-97da-b9d80de56af4-logs" (OuterVolumeSpecName: "logs") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.556998 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-scripts" (OuterVolumeSpecName: "scripts") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.567532 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.585248 4843 scope.go:117] "RemoveContainer" containerID="874949fbd1e7ec53ac8466b5b42f25cb17de84f553f45b3fd652d590b732b24d" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.590969 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0b8901d-9037-4d8e-97da-b9d80de56af4-kube-api-access-rp2dg" (OuterVolumeSpecName: "kube-api-access-rp2dg") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "kube-api-access-rp2dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.598863 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.603688 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.613361 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.630736 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631047 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="cinder-scheduler" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631059 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="cinder-scheduler" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631069 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="probe" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631075 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="probe" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631088 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api-log" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631096 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api-log" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631104 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="proxy-httpd" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631109 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="proxy-httpd" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631146 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-notification-agent" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631152 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-notification-agent" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631162 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-central-agent" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631168 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-central-agent" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631178 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" containerName="mariadb-account-delete" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631184 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" containerName="mariadb-account-delete" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631193 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631199 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631209 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="sg-core" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631214 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="sg-core" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631221 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="cinder-backup" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631228 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="cinder-backup" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.631240 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="probe" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631246 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="probe" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631385 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" containerName="mariadb-account-delete" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631395 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-central-agent" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631401 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api-log" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631409 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="probe" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631417 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="ceilometer-notification-agent" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631424 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="probe" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631455 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" containerName="cinder-api" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631466 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="sg-core" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631472 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" containerName="proxy-httpd" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631485 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="902898bd-0b20-4fd0-991a-58c64f9b3aa0" containerName="cinder-backup" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.631516 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab00353-e46e-455f-98df-a89f1192edc0" containerName="cinder-scheduler" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.633917 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp2dg\" (UniqueName: \"kubernetes.io/projected/d0b8901d-9037-4d8e-97da-b9d80de56af4-kube-api-access-rp2dg\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.633944 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0b8901d-9037-4d8e-97da-b9d80de56af4-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.633972 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.633985 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.633994 4843 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.634001 4843 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0b8901d-9037-4d8e-97da-b9d80de56af4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.635482 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.664522 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.664744 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.665024 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.669335 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data" (OuterVolumeSpecName: "config-data") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.677781 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.683544 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.692216 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.692483 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "d0b8901d-9037-4d8e-97da-b9d80de56af4" (UID: "d0b8901d-9037-4d8e-97da-b9d80de56af4"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.714259 4843 scope.go:117] "RemoveContainer" containerID="edfe00d1ad21ae85435ae3aae021d9a8d5f52231f61849e008f63c0d1ad34db6" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735126 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-log-httpd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735428 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735551 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-config-data\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735678 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-run-httpd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735763 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735841 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.735916 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq4qd\" (UniqueName: \"kubernetes.io/projected/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-kube-api-access-tq4qd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.736003 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-scripts\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.736146 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.736210 4843 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.736320 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.736381 4843 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b8901d-9037-4d8e-97da-b9d80de56af4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.737717 4843 scope.go:117] "RemoveContainer" containerID="3343f6f7dbf0e742dee958b6e6276c71a7074d73a125ecec3c9a21c4f3ef71d4" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.755935 4843 scope.go:117] "RemoveContainer" containerID="535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.761431 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.776420 4843 scope.go:117] "RemoveContainer" containerID="3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.791586 4843 scope.go:117] "RemoveContainer" containerID="535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.792309 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e\": container with ID starting with 535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e not found: ID does not exist" containerID="535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.792402 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e"} err="failed to get container status \"535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e\": rpc error: code = NotFound desc = could not find container \"535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e\": container with ID starting with 535f10025111c789d494340fd21bef13956fb5bca9555add6b37647aa9d4cc5e not found: ID does not exist" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.792457 4843 scope.go:117] "RemoveContainer" containerID="3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f" Mar 14 09:54:59 crc kubenswrapper[4843]: E0314 09:54:59.793015 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f\": container with ID starting with 3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f not found: ID does not exist" containerID="3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.793041 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f"} err="failed to get container status \"3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f\": rpc error: code = NotFound desc = could not find container \"3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f\": container with ID starting with 3083f14030f1140e85ae3d0dfd88db5553ede31e6c4f9c46c104303aa2c2c82f not found: ID does not exist" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.815209 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.823133 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-api-0"] Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837288 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-run-httpd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837439 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837543 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837631 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq4qd\" (UniqueName: \"kubernetes.io/projected/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-kube-api-access-tq4qd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837709 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-scripts\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837806 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-log-httpd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837889 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837980 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-config-data\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.838008 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-log-httpd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.837810 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-run-httpd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.842079 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.842176 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-scripts\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.842265 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-config-data\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.843033 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.844037 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:54:59 crc kubenswrapper[4843]: I0314 09:54:59.858423 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq4qd\" (UniqueName: \"kubernetes.io/projected/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-kube-api-access-tq4qd\") pod \"ceilometer-0\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.011726 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.144840 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-db-create-kwwhw"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.153874 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-db-create-kwwhw"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.162392 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder-9191-account-create-update-hwvmk"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.168313 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/cinder9191-account-delete-4xsz9"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.176110 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder-9191-account-create-update-hwvmk"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.184906 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/cinder9191-account-delete-4xsz9"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.498034 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:00 crc kubenswrapper[4843]: W0314 09:55:00.515037 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a49e7e4_ad73_43d3_8755_8d4f31790b8e.slice/crio-a801865df42a3262400ffb0839d9a292e7cb5a82a0fbb07c03e5e1fe7c3f43be WatchSource:0}: Error finding container a801865df42a3262400ffb0839d9a292e7cb5a82a0fbb07c03e5e1fe7c3f43be: Status 404 returned error can't find the container with id a801865df42a3262400ffb0839d9a292e7cb5a82a0fbb07c03e5e1fe7c3f43be Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.710231 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4vmr"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.712172 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.724774 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4vmr"] Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.752109 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkjbz\" (UniqueName: \"kubernetes.io/projected/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-kube-api-access-wkjbz\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.752173 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-utilities\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.752255 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-catalog-content\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.853712 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkjbz\" (UniqueName: \"kubernetes.io/projected/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-kube-api-access-wkjbz\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.854022 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-utilities\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.854095 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-catalog-content\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.854505 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-utilities\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.854545 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-catalog-content\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.873592 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkjbz\" (UniqueName: \"kubernetes.io/projected/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-kube-api-access-wkjbz\") pod \"community-operators-p4vmr\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:00 crc kubenswrapper[4843]: I0314 09:55:00.921419 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.072612 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.350040 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2db0e8e9-cb24-4896-bc66-31cc71d553cd" path="/var/lib/kubelet/pods/2db0e8e9-cb24-4896-bc66-31cc71d553cd/volumes" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.351025 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60eeb6e3-bcc2-4c90-b551-0db727abcb1d" path="/var/lib/kubelet/pods/60eeb6e3-bcc2-4c90-b551-0db727abcb1d/volumes" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.351469 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8885224-4aed-4a40-8dfa-89b64366597f" path="/var/lib/kubelet/pods/c8885224-4aed-4a40-8dfa-89b64366597f/volumes" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.355924 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0b8901d-9037-4d8e-97da-b9d80de56af4" path="/var/lib/kubelet/pods/d0b8901d-9037-4d8e-97da-b9d80de56af4/volumes" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.356543 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde" path="/var/lib/kubelet/pods/f9127b5f-f6cd-4c5f-ba2d-43cdd29e0dde/volumes" Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.523165 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerStarted","Data":"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd"} Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.523210 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerStarted","Data":"a801865df42a3262400ffb0839d9a292e7cb5a82a0fbb07c03e5e1fe7c3f43be"} Mar 14 09:55:01 crc kubenswrapper[4843]: I0314 09:55:01.659595 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4vmr"] Mar 14 09:55:02 crc kubenswrapper[4843]: I0314 09:55:02.147013 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:55:02 crc kubenswrapper[4843]: I0314 09:55:02.550861 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerStarted","Data":"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9"} Mar 14 09:55:02 crc kubenswrapper[4843]: I0314 09:55:02.550933 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerStarted","Data":"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5"} Mar 14 09:55:02 crc kubenswrapper[4843]: I0314 09:55:02.553032 4843 generic.go:334] "Generic (PLEG): container finished" podID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerID="5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856" exitCode=0 Mar 14 09:55:02 crc kubenswrapper[4843]: I0314 09:55:02.553076 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4vmr" event={"ID":"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9","Type":"ContainerDied","Data":"5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856"} Mar 14 09:55:02 crc kubenswrapper[4843]: I0314 09:55:02.553101 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4vmr" event={"ID":"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9","Type":"ContainerStarted","Data":"19332238e99975bd64fc88e4e3a90e32fc7ed33871d4fc6791b3a3e6de7d8784"} Mar 14 09:55:03 crc kubenswrapper[4843]: I0314 09:55:03.375931 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:55:04 crc kubenswrapper[4843]: I0314 09:55:04.581518 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerStarted","Data":"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7"} Mar 14 09:55:04 crc kubenswrapper[4843]: I0314 09:55:04.581848 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:04 crc kubenswrapper[4843]: I0314 09:55:04.585297 4843 generic.go:334] "Generic (PLEG): container finished" podID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerID="8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4" exitCode=0 Mar 14 09:55:04 crc kubenswrapper[4843]: I0314 09:55:04.585345 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4vmr" event={"ID":"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9","Type":"ContainerDied","Data":"8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4"} Mar 14 09:55:04 crc kubenswrapper[4843]: I0314 09:55:04.600840 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:55:04 crc kubenswrapper[4843]: I0314 09:55:04.612970 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=1.8672907909999998 podStartE2EDuration="5.612950924s" podCreationTimestamp="2026-03-14 09:54:59 +0000 UTC" firstStartedPulling="2026-03-14 09:55:00.517342332 +0000 UTC m=+2667.829953460" lastFinishedPulling="2026-03-14 09:55:04.263002465 +0000 UTC m=+2671.575613593" observedRunningTime="2026-03-14 09:55:04.612231227 +0000 UTC m=+2671.924842355" watchObservedRunningTime="2026-03-14 09:55:04.612950924 +0000 UTC m=+2671.925562072" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.514912 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j8b6n"] Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.518099 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.529647 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j8b6n"] Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.532682 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-catalog-content\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.532720 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nskfb\" (UniqueName: \"kubernetes.io/projected/4e0036f2-535c-47e5-b145-83c38ea91d71-kube-api-access-nskfb\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.532774 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-utilities\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.607340 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4vmr" event={"ID":"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9","Type":"ContainerStarted","Data":"bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1"} Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.635152 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4vmr" podStartSLOduration=3.154203913 podStartE2EDuration="5.635138775s" podCreationTimestamp="2026-03-14 09:55:00 +0000 UTC" firstStartedPulling="2026-03-14 09:55:02.55458515 +0000 UTC m=+2669.867196278" lastFinishedPulling="2026-03-14 09:55:05.035520012 +0000 UTC m=+2672.348131140" observedRunningTime="2026-03-14 09:55:05.631266721 +0000 UTC m=+2672.943877849" watchObservedRunningTime="2026-03-14 09:55:05.635138775 +0000 UTC m=+2672.947749903" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.641244 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-catalog-content\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.641318 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nskfb\" (UniqueName: \"kubernetes.io/projected/4e0036f2-535c-47e5-b145-83c38ea91d71-kube-api-access-nskfb\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.641413 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-utilities\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.642757 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-utilities\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.642974 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-catalog-content\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.666862 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nskfb\" (UniqueName: \"kubernetes.io/projected/4e0036f2-535c-47e5-b145-83c38ea91d71-kube-api-access-nskfb\") pod \"redhat-operators-j8b6n\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.838607 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:05 crc kubenswrapper[4843]: I0314 09:55:05.843848 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_b1ef26b5-077f-4419-87a4-b1100c18f189/watcher-decision-engine/0.log" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.168467 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254315 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-cert-memcached-mtls\") pod \"b1ef26b5-077f-4419-87a4-b1100c18f189\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254390 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ef26b5-077f-4419-87a4-b1100c18f189-logs\") pod \"b1ef26b5-077f-4419-87a4-b1100c18f189\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254409 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxlsq\" (UniqueName: \"kubernetes.io/projected/b1ef26b5-077f-4419-87a4-b1100c18f189-kube-api-access-kxlsq\") pod \"b1ef26b5-077f-4419-87a4-b1100c18f189\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254463 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-combined-ca-bundle\") pod \"b1ef26b5-077f-4419-87a4-b1100c18f189\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254514 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-custom-prometheus-ca\") pod \"b1ef26b5-077f-4419-87a4-b1100c18f189\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254602 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-config-data\") pod \"b1ef26b5-077f-4419-87a4-b1100c18f189\" (UID: \"b1ef26b5-077f-4419-87a4-b1100c18f189\") " Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254721 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1ef26b5-077f-4419-87a4-b1100c18f189-logs" (OuterVolumeSpecName: "logs") pod "b1ef26b5-077f-4419-87a4-b1100c18f189" (UID: "b1ef26b5-077f-4419-87a4-b1100c18f189"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.254869 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ef26b5-077f-4419-87a4-b1100c18f189-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.260285 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ef26b5-077f-4419-87a4-b1100c18f189-kube-api-access-kxlsq" (OuterVolumeSpecName: "kube-api-access-kxlsq") pod "b1ef26b5-077f-4419-87a4-b1100c18f189" (UID: "b1ef26b5-077f-4419-87a4-b1100c18f189"). InnerVolumeSpecName "kube-api-access-kxlsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.282244 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "b1ef26b5-077f-4419-87a4-b1100c18f189" (UID: "b1ef26b5-077f-4419-87a4-b1100c18f189"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.287914 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1ef26b5-077f-4419-87a4-b1100c18f189" (UID: "b1ef26b5-077f-4419-87a4-b1100c18f189"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.303579 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-config-data" (OuterVolumeSpecName: "config-data") pod "b1ef26b5-077f-4419-87a4-b1100c18f189" (UID: "b1ef26b5-077f-4419-87a4-b1100c18f189"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.332413 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "b1ef26b5-077f-4419-87a4-b1100c18f189" (UID: "b1ef26b5-077f-4419-87a4-b1100c18f189"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.356404 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.356441 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.356453 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxlsq\" (UniqueName: \"kubernetes.io/projected/b1ef26b5-077f-4419-87a4-b1100c18f189-kube-api-access-kxlsq\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.356463 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.356473 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/b1ef26b5-077f-4419-87a4-b1100c18f189-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.363045 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j8b6n"] Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.616908 4843 generic.go:334] "Generic (PLEG): container finished" podID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerID="da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109" exitCode=0 Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.617844 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerDied","Data":"da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109"} Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.621315 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerStarted","Data":"5b3bde5b6a63c969b1462d69f533b4251aa5f69e936390c85ceb01502d290f8a"} Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.621422 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"b1ef26b5-077f-4419-87a4-b1100c18f189","Type":"ContainerDied","Data":"80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862"} Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.621521 4843 scope.go:117] "RemoveContainer" containerID="80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.620336 4843 generic.go:334] "Generic (PLEG): container finished" podID="b1ef26b5-077f-4419-87a4-b1100c18f189" containerID="80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862" exitCode=0 Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.621730 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"b1ef26b5-077f-4419-87a4-b1100c18f189","Type":"ContainerDied","Data":"df975bdda7da46be44e7b401c0eab148480a374817a425b05befc9bef43d1cad"} Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.620400 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.650316 4843 scope.go:117] "RemoveContainer" containerID="80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862" Mar 14 09:55:06 crc kubenswrapper[4843]: E0314 09:55:06.652737 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862\": container with ID starting with 80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862 not found: ID does not exist" containerID="80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.652917 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862"} err="failed to get container status \"80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862\": rpc error: code = NotFound desc = could not find container \"80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862\": container with ID starting with 80b07b8d5c7e191c6758ec3420efcb3fb38861919f0cbbad6ace01d48a929862 not found: ID does not exist" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.670347 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.679484 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.689750 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:06 crc kubenswrapper[4843]: E0314 09:55:06.690104 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ef26b5-077f-4419-87a4-b1100c18f189" containerName="watcher-decision-engine" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.690120 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ef26b5-077f-4419-87a4-b1100c18f189" containerName="watcher-decision-engine" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.690331 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ef26b5-077f-4419-87a4-b1100c18f189" containerName="watcher-decision-engine" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.690947 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.693347 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.711756 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.766171 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.766220 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.766291 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.766311 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.766366 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.766435 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtwdl\" (UniqueName: \"kubernetes.io/projected/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-kube-api-access-qtwdl\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.867532 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.867607 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtwdl\" (UniqueName: \"kubernetes.io/projected/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-kube-api-access-qtwdl\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.867653 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.867678 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.867720 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.867736 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.868201 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.873043 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.878798 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.882822 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.882909 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:06 crc kubenswrapper[4843]: I0314 09:55:06.886440 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtwdl\" (UniqueName: \"kubernetes.io/projected/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-kube-api-access-qtwdl\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:07 crc kubenswrapper[4843]: I0314 09:55:07.037195 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:07 crc kubenswrapper[4843]: I0314 09:55:07.357809 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1ef26b5-077f-4419-87a4-b1100c18f189" path="/var/lib/kubelet/pods/b1ef26b5-077f-4419-87a4-b1100c18f189/volumes" Mar 14 09:55:07 crc kubenswrapper[4843]: I0314 09:55:07.506991 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:07 crc kubenswrapper[4843]: W0314 09:55:07.509899 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e98ba2a_93b1_47d4_8e7e_dddb3ca56fc4.slice/crio-ea5623acbb3ae60cc43c3ef0c979cd0763b24118a241150777f089425ef4e0da WatchSource:0}: Error finding container ea5623acbb3ae60cc43c3ef0c979cd0763b24118a241150777f089425ef4e0da: Status 404 returned error can't find the container with id ea5623acbb3ae60cc43c3ef0c979cd0763b24118a241150777f089425ef4e0da Mar 14 09:55:07 crc kubenswrapper[4843]: I0314 09:55:07.631453 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4","Type":"ContainerStarted","Data":"ea5623acbb3ae60cc43c3ef0c979cd0763b24118a241150777f089425ef4e0da"} Mar 14 09:55:07 crc kubenswrapper[4843]: I0314 09:55:07.636014 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerStarted","Data":"a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f"} Mar 14 09:55:08 crc kubenswrapper[4843]: I0314 09:55:08.657143 4843 generic.go:334] "Generic (PLEG): container finished" podID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerID="a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f" exitCode=0 Mar 14 09:55:08 crc kubenswrapper[4843]: I0314 09:55:08.657258 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerDied","Data":"a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f"} Mar 14 09:55:08 crc kubenswrapper[4843]: I0314 09:55:08.664103 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4","Type":"ContainerStarted","Data":"e4f681c58540489ce55e7cd8f6a8b0ddb618c031e8d78d1c1ac328835e99a228"} Mar 14 09:55:08 crc kubenswrapper[4843]: I0314 09:55:08.715332 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.715311681 podStartE2EDuration="2.715311681s" podCreationTimestamp="2026-03-14 09:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:55:08.710656077 +0000 UTC m=+2676.023267205" watchObservedRunningTime="2026-03-14 09:55:08.715311681 +0000 UTC m=+2676.027922809" Mar 14 09:55:09 crc kubenswrapper[4843]: I0314 09:55:09.378945 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:09 crc kubenswrapper[4843]: I0314 09:55:09.675075 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerStarted","Data":"473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96"} Mar 14 09:55:09 crc kubenswrapper[4843]: I0314 09:55:09.715310 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j8b6n" podStartSLOduration=2.227111259 podStartE2EDuration="4.715291189s" podCreationTimestamp="2026-03-14 09:55:05 +0000 UTC" firstStartedPulling="2026-03-14 09:55:06.618528636 +0000 UTC m=+2673.931139764" lastFinishedPulling="2026-03-14 09:55:09.106708556 +0000 UTC m=+2676.419319694" observedRunningTime="2026-03-14 09:55:09.708604665 +0000 UTC m=+2677.021215813" watchObservedRunningTime="2026-03-14 09:55:09.715291189 +0000 UTC m=+2677.027902327" Mar 14 09:55:10 crc kubenswrapper[4843]: I0314 09:55:10.610259 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:11 crc kubenswrapper[4843]: I0314 09:55:11.073612 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:11 crc kubenswrapper[4843]: I0314 09:55:11.073671 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:11 crc kubenswrapper[4843]: I0314 09:55:11.117684 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:11 crc kubenswrapper[4843]: I0314 09:55:11.743923 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:11 crc kubenswrapper[4843]: I0314 09:55:11.788789 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:12 crc kubenswrapper[4843]: I0314 09:55:12.995494 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:13 crc kubenswrapper[4843]: I0314 09:55:13.102142 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4vmr"] Mar 14 09:55:13 crc kubenswrapper[4843]: I0314 09:55:13.705129 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4vmr" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="registry-server" containerID="cri-o://bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1" gracePeriod=2 Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.168960 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.188691 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.196721 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-catalog-content\") pod \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.196843 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-utilities\") pod \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.196994 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkjbz\" (UniqueName: \"kubernetes.io/projected/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-kube-api-access-wkjbz\") pod \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\" (UID: \"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9\") " Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.199596 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-utilities" (OuterVolumeSpecName: "utilities") pod "e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" (UID: "e7e4843f-0ee3-4643-8c0e-1d876d1b41f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.206495 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-kube-api-access-wkjbz" (OuterVolumeSpecName: "kube-api-access-wkjbz") pod "e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" (UID: "e7e4843f-0ee3-4643-8c0e-1d876d1b41f9"). InnerVolumeSpecName "kube-api-access-wkjbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.298958 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.298995 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkjbz\" (UniqueName: \"kubernetes.io/projected/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-kube-api-access-wkjbz\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.726206 4843 generic.go:334] "Generic (PLEG): container finished" podID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerID="bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1" exitCode=0 Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.726253 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4vmr" event={"ID":"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9","Type":"ContainerDied","Data":"bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1"} Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.726295 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4vmr" event={"ID":"e7e4843f-0ee3-4643-8c0e-1d876d1b41f9","Type":"ContainerDied","Data":"19332238e99975bd64fc88e4e3a90e32fc7ed33871d4fc6791b3a3e6de7d8784"} Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.726313 4843 scope.go:117] "RemoveContainer" containerID="bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.726436 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4vmr" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.749196 4843 scope.go:117] "RemoveContainer" containerID="8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.769132 4843 scope.go:117] "RemoveContainer" containerID="5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.779183 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" (UID: "e7e4843f-0ee3-4643-8c0e-1d876d1b41f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.806515 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.808229 4843 scope.go:117] "RemoveContainer" containerID="bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1" Mar 14 09:55:14 crc kubenswrapper[4843]: E0314 09:55:14.808743 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1\": container with ID starting with bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1 not found: ID does not exist" containerID="bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.808799 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1"} err="failed to get container status \"bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1\": rpc error: code = NotFound desc = could not find container \"bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1\": container with ID starting with bb5d24da8cf51903f8fcebffd8666644db612b964332cf82a8b2ac8251804ae1 not found: ID does not exist" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.808829 4843 scope.go:117] "RemoveContainer" containerID="8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4" Mar 14 09:55:14 crc kubenswrapper[4843]: E0314 09:55:14.809434 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4\": container with ID starting with 8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4 not found: ID does not exist" containerID="8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.809478 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4"} err="failed to get container status \"8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4\": rpc error: code = NotFound desc = could not find container \"8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4\": container with ID starting with 8473a7a5edde2a63856b70a6fc09ac093a0f57cb71feaf08af53d3fcc7454ad4 not found: ID does not exist" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.809510 4843 scope.go:117] "RemoveContainer" containerID="5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856" Mar 14 09:55:14 crc kubenswrapper[4843]: E0314 09:55:14.809868 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856\": container with ID starting with 5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856 not found: ID does not exist" containerID="5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856" Mar 14 09:55:14 crc kubenswrapper[4843]: I0314 09:55:14.809898 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856"} err="failed to get container status \"5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856\": rpc error: code = NotFound desc = could not find container \"5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856\": container with ID starting with 5dba25fcb5070b93550117615403155e5b11769f16a37d553fdc5cc49f16e856 not found: ID does not exist" Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.054993 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4vmr"] Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.082105 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4vmr"] Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.348456 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" path="/var/lib/kubelet/pods/e7e4843f-0ee3-4643-8c0e-1d876d1b41f9/volumes" Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.367021 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.387576 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.387638 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.839179 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:15 crc kubenswrapper[4843]: I0314 09:55:15.839629 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:16 crc kubenswrapper[4843]: I0314 09:55:16.586406 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:16 crc kubenswrapper[4843]: I0314 09:55:16.884984 4843 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j8b6n" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="registry-server" probeResult="failure" output=< Mar 14 09:55:16 crc kubenswrapper[4843]: timeout: failed to connect service ":50051" within 1s Mar 14 09:55:16 crc kubenswrapper[4843]: > Mar 14 09:55:17 crc kubenswrapper[4843]: I0314 09:55:17.038126 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:17 crc kubenswrapper[4843]: I0314 09:55:17.074874 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:17 crc kubenswrapper[4843]: I0314 09:55:17.752404 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:17 crc kubenswrapper[4843]: I0314 09:55:17.779999 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:17 crc kubenswrapper[4843]: I0314 09:55:17.815709 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.025012 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_watcher-kuttl-decision-engine-0_0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/watcher-decision-engine/0.log" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.155253 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-8skbb"] Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.160833 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-8skbb"] Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.211528 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher1522-account-delete-9vntc"] Mar 14 09:55:19 crc kubenswrapper[4843]: E0314 09:55:19.211927 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="registry-server" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.211943 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="registry-server" Mar 14 09:55:19 crc kubenswrapper[4843]: E0314 09:55:19.211956 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="extract-utilities" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.211963 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="extract-utilities" Mar 14 09:55:19 crc kubenswrapper[4843]: E0314 09:55:19.211984 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="extract-content" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.211991 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="extract-content" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.212145 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e4843f-0ee3-4643-8c0e-1d876d1b41f9" containerName="registry-server" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.212825 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.219294 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.219552 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="53784836-318f-4a01-899d-f3cd6cbd72d0" containerName="watcher-applier" containerID="cri-o://53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4" gracePeriod=30 Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.231343 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher1522-account-delete-9vntc"] Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.281004 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.281219 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-kuttl-api-log" containerID="cri-o://244741a0a9f9acfe64beeb18c8adfd7376b0f149919d85748105035f225f3558" gracePeriod=30 Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.281604 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-api" containerID="cri-o://efe65270a070c09c320a96842e53a64e21ee058c9ab97a7d1731b6135fc64e28" gracePeriod=30 Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.334514 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.352882 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2408d8-efbd-44b0-a7cd-98a5891f7fab" path="/var/lib/kubelet/pods/bc2408d8-efbd-44b0-a7cd-98a5891f7fab/volumes" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.380876 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27b43d13-6378-47a9-b3a9-3e9532e47977-operator-scripts\") pod \"watcher1522-account-delete-9vntc\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.380944 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4slv4\" (UniqueName: \"kubernetes.io/projected/27b43d13-6378-47a9-b3a9-3e9532e47977-kube-api-access-4slv4\") pod \"watcher1522-account-delete-9vntc\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.483210 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27b43d13-6378-47a9-b3a9-3e9532e47977-operator-scripts\") pod \"watcher1522-account-delete-9vntc\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.483297 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4slv4\" (UniqueName: \"kubernetes.io/projected/27b43d13-6378-47a9-b3a9-3e9532e47977-kube-api-access-4slv4\") pod \"watcher1522-account-delete-9vntc\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.484002 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27b43d13-6378-47a9-b3a9-3e9532e47977-operator-scripts\") pod \"watcher1522-account-delete-9vntc\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.507485 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4slv4\" (UniqueName: \"kubernetes.io/projected/27b43d13-6378-47a9-b3a9-3e9532e47977-kube-api-access-4slv4\") pod \"watcher1522-account-delete-9vntc\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.533723 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.778357 4843 generic.go:334] "Generic (PLEG): container finished" podID="d4524e41-3177-4936-b995-6b0231394c16" containerID="244741a0a9f9acfe64beeb18c8adfd7376b0f149919d85748105035f225f3558" exitCode=143 Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.778586 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d4524e41-3177-4936-b995-6b0231394c16","Type":"ContainerDied","Data":"244741a0a9f9acfe64beeb18c8adfd7376b0f149919d85748105035f225f3558"} Mar 14 09:55:19 crc kubenswrapper[4843]: I0314 09:55:19.779075 4843 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" secret="" err="secret \"watcher-watcher-kuttl-dockercfg-dzfjc\" not found" Mar 14 09:55:19 crc kubenswrapper[4843]: E0314 09:55:19.891303 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:19 crc kubenswrapper[4843]: E0314 09:55:19.891401 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data podName:0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4 nodeName:}" failed. No retries permitted until 2026-03-14 09:55:20.391373646 +0000 UTC m=+2687.703984774 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4") : secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.024651 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher1522-account-delete-9vntc"] Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.344937 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:20 crc kubenswrapper[4843]: E0314 09:55:20.399048 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:20 crc kubenswrapper[4843]: E0314 09:55:20.399136 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data podName:0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4 nodeName:}" failed. No retries permitted until 2026-03-14 09:55:21.399117959 +0000 UTC m=+2688.711729087 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4") : secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.499957 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-config-data\") pod \"53784836-318f-4a01-899d-f3cd6cbd72d0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.500009 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-combined-ca-bundle\") pod \"53784836-318f-4a01-899d-f3cd6cbd72d0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.500031 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vctnx\" (UniqueName: \"kubernetes.io/projected/53784836-318f-4a01-899d-f3cd6cbd72d0-kube-api-access-vctnx\") pod \"53784836-318f-4a01-899d-f3cd6cbd72d0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.500108 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-cert-memcached-mtls\") pod \"53784836-318f-4a01-899d-f3cd6cbd72d0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.500128 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53784836-318f-4a01-899d-f3cd6cbd72d0-logs\") pod \"53784836-318f-4a01-899d-f3cd6cbd72d0\" (UID: \"53784836-318f-4a01-899d-f3cd6cbd72d0\") " Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.508135 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53784836-318f-4a01-899d-f3cd6cbd72d0-logs" (OuterVolumeSpecName: "logs") pod "53784836-318f-4a01-899d-f3cd6cbd72d0" (UID: "53784836-318f-4a01-899d-f3cd6cbd72d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.508465 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53784836-318f-4a01-899d-f3cd6cbd72d0-kube-api-access-vctnx" (OuterVolumeSpecName: "kube-api-access-vctnx") pod "53784836-318f-4a01-899d-f3cd6cbd72d0" (UID: "53784836-318f-4a01-899d-f3cd6cbd72d0"). InnerVolumeSpecName "kube-api-access-vctnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.531907 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53784836-318f-4a01-899d-f3cd6cbd72d0" (UID: "53784836-318f-4a01-899d-f3cd6cbd72d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.549170 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-config-data" (OuterVolumeSpecName: "config-data") pod "53784836-318f-4a01-899d-f3cd6cbd72d0" (UID: "53784836-318f-4a01-899d-f3cd6cbd72d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.582566 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "53784836-318f-4a01-899d-f3cd6cbd72d0" (UID: "53784836-318f-4a01-899d-f3cd6cbd72d0"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.601783 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.601817 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.601828 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vctnx\" (UniqueName: \"kubernetes.io/projected/53784836-318f-4a01-899d-f3cd6cbd72d0-kube-api-access-vctnx\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.601837 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/53784836-318f-4a01-899d-f3cd6cbd72d0-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.601845 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53784836-318f-4a01-899d-f3cd6cbd72d0-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.866417 4843 generic.go:334] "Generic (PLEG): container finished" podID="53784836-318f-4a01-899d-f3cd6cbd72d0" containerID="53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4" exitCode=0 Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.867302 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.867394 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"53784836-318f-4a01-899d-f3cd6cbd72d0","Type":"ContainerDied","Data":"53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4"} Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.867465 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"53784836-318f-4a01-899d-f3cd6cbd72d0","Type":"ContainerDied","Data":"e1edbecd83725c596f781b5d784089c89050e6b1b15f23a9e5214826ef76b2f3"} Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.867508 4843 scope.go:117] "RemoveContainer" containerID="53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.877897 4843 generic.go:334] "Generic (PLEG): container finished" podID="d4524e41-3177-4936-b995-6b0231394c16" containerID="efe65270a070c09c320a96842e53a64e21ee058c9ab97a7d1731b6135fc64e28" exitCode=0 Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.877969 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d4524e41-3177-4936-b995-6b0231394c16","Type":"ContainerDied","Data":"efe65270a070c09c320a96842e53a64e21ee058c9ab97a7d1731b6135fc64e28"} Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.884567 4843 generic.go:334] "Generic (PLEG): container finished" podID="27b43d13-6378-47a9-b3a9-3e9532e47977" containerID="6f0119e162497e6f567603380b44c3b32a8cf39558a29296139d301f1d7bff79" exitCode=0 Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.884872 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" containerName="watcher-decision-engine" containerID="cri-o://e4f681c58540489ce55e7cd8f6a8b0ddb618c031e8d78d1c1ac328835e99a228" gracePeriod=30 Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.885336 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" event={"ID":"27b43d13-6378-47a9-b3a9-3e9532e47977","Type":"ContainerDied","Data":"6f0119e162497e6f567603380b44c3b32a8cf39558a29296139d301f1d7bff79"} Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.885362 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" event={"ID":"27b43d13-6378-47a9-b3a9-3e9532e47977","Type":"ContainerStarted","Data":"33be2eaadd48ab9b1ad4e7739a7fb61ab14319daa96bf85a527069ea2d8c701b"} Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.936357 4843 scope.go:117] "RemoveContainer" containerID="53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4" Mar 14 09:55:20 crc kubenswrapper[4843]: E0314 09:55:20.941056 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4\": container with ID starting with 53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4 not found: ID does not exist" containerID="53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.941088 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4"} err="failed to get container status \"53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4\": rpc error: code = NotFound desc = could not find container \"53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4\": container with ID starting with 53042393c8a79b53c58af28a279b18d898e3e7bf49d32e4f91010da69bef99f4 not found: ID does not exist" Mar 14 09:55:20 crc kubenswrapper[4843]: I0314 09:55:20.999332 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.024651 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.040318 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.151996 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-cert-memcached-mtls\") pod \"d4524e41-3177-4936-b995-6b0231394c16\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.152410 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-combined-ca-bundle\") pod \"d4524e41-3177-4936-b995-6b0231394c16\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.152469 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4524e41-3177-4936-b995-6b0231394c16-logs\") pod \"d4524e41-3177-4936-b995-6b0231394c16\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.152488 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-custom-prometheus-ca\") pod \"d4524e41-3177-4936-b995-6b0231394c16\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.152507 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftxbv\" (UniqueName: \"kubernetes.io/projected/d4524e41-3177-4936-b995-6b0231394c16-kube-api-access-ftxbv\") pod \"d4524e41-3177-4936-b995-6b0231394c16\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.152648 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-config-data\") pod \"d4524e41-3177-4936-b995-6b0231394c16\" (UID: \"d4524e41-3177-4936-b995-6b0231394c16\") " Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.152965 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4524e41-3177-4936-b995-6b0231394c16-logs" (OuterVolumeSpecName: "logs") pod "d4524e41-3177-4936-b995-6b0231394c16" (UID: "d4524e41-3177-4936-b995-6b0231394c16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.157085 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4524e41-3177-4936-b995-6b0231394c16-kube-api-access-ftxbv" (OuterVolumeSpecName: "kube-api-access-ftxbv") pod "d4524e41-3177-4936-b995-6b0231394c16" (UID: "d4524e41-3177-4936-b995-6b0231394c16"). InnerVolumeSpecName "kube-api-access-ftxbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.183652 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4524e41-3177-4936-b995-6b0231394c16" (UID: "d4524e41-3177-4936-b995-6b0231394c16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.196435 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "d4524e41-3177-4936-b995-6b0231394c16" (UID: "d4524e41-3177-4936-b995-6b0231394c16"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.209192 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-config-data" (OuterVolumeSpecName: "config-data") pod "d4524e41-3177-4936-b995-6b0231394c16" (UID: "d4524e41-3177-4936-b995-6b0231394c16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.223742 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "d4524e41-3177-4936-b995-6b0231394c16" (UID: "d4524e41-3177-4936-b995-6b0231394c16"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.254537 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.254570 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.254581 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.254593 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4524e41-3177-4936-b995-6b0231394c16-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.254601 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d4524e41-3177-4936-b995-6b0231394c16-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.254610 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftxbv\" (UniqueName: \"kubernetes.io/projected/d4524e41-3177-4936-b995-6b0231394c16-kube-api-access-ftxbv\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.349116 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53784836-318f-4a01-899d-f3cd6cbd72d0" path="/var/lib/kubelet/pods/53784836-318f-4a01-899d-f3cd6cbd72d0/volumes" Mar 14 09:55:21 crc kubenswrapper[4843]: E0314 09:55:21.458119 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:21 crc kubenswrapper[4843]: E0314 09:55:21.458182 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data podName:0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4 nodeName:}" failed. No retries permitted until 2026-03-14 09:55:23.458164643 +0000 UTC m=+2690.770775771 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4") : secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.761409 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.761745 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-central-agent" containerID="cri-o://65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd" gracePeriod=30 Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.761809 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="sg-core" containerID="cri-o://ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9" gracePeriod=30 Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.761852 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-notification-agent" containerID="cri-o://98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5" gracePeriod=30 Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.761951 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="proxy-httpd" containerID="cri-o://4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7" gracePeriod=30 Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.862218 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.2:3000/\": read tcp 10.217.0.2:48178->10.217.1.2:3000: read: connection reset by peer" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.908350 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"d4524e41-3177-4936-b995-6b0231394c16","Type":"ContainerDied","Data":"d3de27c941bc73b9effafc1c2d7b8d8e6adf8b335c4542126549ede4772a3b57"} Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.908406 4843 scope.go:117] "RemoveContainer" containerID="efe65270a070c09c320a96842e53a64e21ee058c9ab97a7d1731b6135fc64e28" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.908530 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.922983 4843 generic.go:334] "Generic (PLEG): container finished" podID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerID="4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7" exitCode=0 Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.923264 4843 generic.go:334] "Generic (PLEG): container finished" podID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerID="ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9" exitCode=2 Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.923081 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerDied","Data":"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7"} Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.923483 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerDied","Data":"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9"} Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.966673 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.984511 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:55:21 crc kubenswrapper[4843]: I0314 09:55:21.986163 4843 scope.go:117] "RemoveContainer" containerID="244741a0a9f9acfe64beeb18c8adfd7376b0f149919d85748105035f225f3558" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.297173 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.377253 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4slv4\" (UniqueName: \"kubernetes.io/projected/27b43d13-6378-47a9-b3a9-3e9532e47977-kube-api-access-4slv4\") pod \"27b43d13-6378-47a9-b3a9-3e9532e47977\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.377328 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27b43d13-6378-47a9-b3a9-3e9532e47977-operator-scripts\") pod \"27b43d13-6378-47a9-b3a9-3e9532e47977\" (UID: \"27b43d13-6378-47a9-b3a9-3e9532e47977\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.378100 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27b43d13-6378-47a9-b3a9-3e9532e47977-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27b43d13-6378-47a9-b3a9-3e9532e47977" (UID: "27b43d13-6378-47a9-b3a9-3e9532e47977"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.382826 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27b43d13-6378-47a9-b3a9-3e9532e47977-kube-api-access-4slv4" (OuterVolumeSpecName: "kube-api-access-4slv4") pod "27b43d13-6378-47a9-b3a9-3e9532e47977" (UID: "27b43d13-6378-47a9-b3a9-3e9532e47977"). InnerVolumeSpecName "kube-api-access-4slv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.479610 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4slv4\" (UniqueName: \"kubernetes.io/projected/27b43d13-6378-47a9-b3a9-3e9532e47977-kube-api-access-4slv4\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.479893 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27b43d13-6378-47a9-b3a9-3e9532e47977-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.599644 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.681885 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-config-data\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.681977 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-sg-core-conf-yaml\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682085 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq4qd\" (UniqueName: \"kubernetes.io/projected/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-kube-api-access-tq4qd\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682117 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-combined-ca-bundle\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682148 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-ceilometer-tls-certs\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682223 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-log-httpd\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682363 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-scripts\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682406 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-run-httpd\") pod \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\" (UID: \"8a49e7e4-ad73-43d3-8755-8d4f31790b8e\") " Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.682982 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.683213 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.696599 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-kube-api-access-tq4qd" (OuterVolumeSpecName: "kube-api-access-tq4qd") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "kube-api-access-tq4qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.697776 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-scripts" (OuterVolumeSpecName: "scripts") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.709630 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.734441 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.753559 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784467 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784498 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq4qd\" (UniqueName: \"kubernetes.io/projected/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-kube-api-access-tq4qd\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784511 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784521 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784531 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784542 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.784554 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.794080 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-config-data" (OuterVolumeSpecName: "config-data") pod "8a49e7e4-ad73-43d3-8755-8d4f31790b8e" (UID: "8a49e7e4-ad73-43d3-8755-8d4f31790b8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.886148 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a49e7e4-ad73-43d3-8755-8d4f31790b8e-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934513 4843 generic.go:334] "Generic (PLEG): container finished" podID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerID="98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5" exitCode=0 Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934544 4843 generic.go:334] "Generic (PLEG): container finished" podID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerID="65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd" exitCode=0 Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934593 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerDied","Data":"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5"} Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934611 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934634 4843 scope.go:117] "RemoveContainer" containerID="4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934622 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerDied","Data":"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd"} Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.934817 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"8a49e7e4-ad73-43d3-8755-8d4f31790b8e","Type":"ContainerDied","Data":"a801865df42a3262400ffb0839d9a292e7cb5a82a0fbb07c03e5e1fe7c3f43be"} Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.938690 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" event={"ID":"27b43d13-6378-47a9-b3a9-3e9532e47977","Type":"ContainerDied","Data":"33be2eaadd48ab9b1ad4e7739a7fb61ab14319daa96bf85a527069ea2d8c701b"} Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.938728 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33be2eaadd48ab9b1ad4e7739a7fb61ab14319daa96bf85a527069ea2d8c701b" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.938746 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher1522-account-delete-9vntc" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.965194 4843 scope.go:117] "RemoveContainer" containerID="ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9" Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.984400 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.990006 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:22 crc kubenswrapper[4843]: I0314 09:55:22.996024 4843 scope.go:117] "RemoveContainer" containerID="98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.009740 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010194 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="proxy-httpd" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010231 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="proxy-httpd" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010255 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-api" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010293 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-api" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010322 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="sg-core" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010333 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="sg-core" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010347 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-kuttl-api-log" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010358 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-kuttl-api-log" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010375 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-notification-agent" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010387 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-notification-agent" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010408 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27b43d13-6378-47a9-b3a9-3e9532e47977" containerName="mariadb-account-delete" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010419 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="27b43d13-6378-47a9-b3a9-3e9532e47977" containerName="mariadb-account-delete" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010458 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53784836-318f-4a01-899d-f3cd6cbd72d0" containerName="watcher-applier" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010470 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="53784836-318f-4a01-899d-f3cd6cbd72d0" containerName="watcher-applier" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.010490 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-central-agent" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.010502 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-central-agent" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.018714 4843 scope.go:117] "RemoveContainer" containerID="65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020702 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="53784836-318f-4a01-899d-f3cd6cbd72d0" containerName="watcher-applier" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020743 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-notification-agent" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020772 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="proxy-httpd" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020786 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="ceilometer-central-agent" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020805 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-kuttl-api-log" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020822 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="27b43d13-6378-47a9-b3a9-3e9532e47977" containerName="mariadb-account-delete" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020850 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4524e41-3177-4936-b995-6b0231394c16" containerName="watcher-api" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.020863 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" containerName="sg-core" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.023234 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.023392 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.026280 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.026405 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.026436 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.053716 4843 scope.go:117] "RemoveContainer" containerID="4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.054837 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7\": container with ID starting with 4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7 not found: ID does not exist" containerID="4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.054914 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7"} err="failed to get container status \"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7\": rpc error: code = NotFound desc = could not find container \"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7\": container with ID starting with 4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7 not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.054957 4843 scope.go:117] "RemoveContainer" containerID="ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.058390 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9\": container with ID starting with ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9 not found: ID does not exist" containerID="ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.058435 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9"} err="failed to get container status \"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9\": rpc error: code = NotFound desc = could not find container \"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9\": container with ID starting with ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9 not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.058465 4843 scope.go:117] "RemoveContainer" containerID="98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.059266 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5\": container with ID starting with 98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5 not found: ID does not exist" containerID="98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.059428 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5"} err="failed to get container status \"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5\": rpc error: code = NotFound desc = could not find container \"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5\": container with ID starting with 98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5 not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.059457 4843 scope.go:117] "RemoveContainer" containerID="65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.059742 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd\": container with ID starting with 65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd not found: ID does not exist" containerID="65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.059778 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd"} err="failed to get container status \"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd\": rpc error: code = NotFound desc = could not find container \"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd\": container with ID starting with 65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.059812 4843 scope.go:117] "RemoveContainer" containerID="4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.060234 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7"} err="failed to get container status \"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7\": rpc error: code = NotFound desc = could not find container \"4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7\": container with ID starting with 4a821516b06f331cf266eebc1d6ae75804b99e41272edb7be9f4dbaea0ca84f7 not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.060304 4843 scope.go:117] "RemoveContainer" containerID="ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.060675 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9"} err="failed to get container status \"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9\": rpc error: code = NotFound desc = could not find container \"ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9\": container with ID starting with ed326b639c354b5628c75cc0aa93792b7d54e2e440642fd67b8d50ff4a43b3e9 not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.060701 4843 scope.go:117] "RemoveContainer" containerID="98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.061064 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5"} err="failed to get container status \"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5\": rpc error: code = NotFound desc = could not find container \"98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5\": container with ID starting with 98811f6b64aa45d78f905861fc7adb85dee5e3a30edec39683cc428bb0d973a5 not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.061089 4843 scope.go:117] "RemoveContainer" containerID="65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.061263 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd"} err="failed to get container status \"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd\": rpc error: code = NotFound desc = could not find container \"65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd\": container with ID starting with 65b62378671cdc7ae7bc3185038ff3a6ba0228f66f28daaabe1da6a3a51377bd not found: ID does not exist" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089295 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089357 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q5jg\" (UniqueName: \"kubernetes.io/projected/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-kube-api-access-7q5jg\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089403 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089551 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-log-httpd\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089608 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-config-data\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089801 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-scripts\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089835 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-run-httpd\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.089874 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.191669 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.191972 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q5jg\" (UniqueName: \"kubernetes.io/projected/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-kube-api-access-7q5jg\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192004 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192033 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-log-httpd\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192053 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-config-data\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192126 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-scripts\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192143 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-run-httpd\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192164 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192764 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-log-httpd\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.192806 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-run-httpd\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.197234 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.197482 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.198764 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-scripts\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.198829 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-config-data\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.201979 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.207484 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q5jg\" (UniqueName: \"kubernetes.io/projected/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-kube-api-access-7q5jg\") pod \"ceilometer-0\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.352169 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a49e7e4-ad73-43d3-8755-8d4f31790b8e" path="/var/lib/kubelet/pods/8a49e7e4-ad73-43d3-8755-8d4f31790b8e/volumes" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.353927 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4524e41-3177-4936-b995-6b0231394c16" path="/var/lib/kubelet/pods/d4524e41-3177-4936-b995-6b0231394c16/volumes" Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.356687 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.496060 4843 secret.go:188] Couldn't get secret watcher-kuttl-default/watcher-kuttl-decision-engine-config-data: secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:23 crc kubenswrapper[4843]: E0314 09:55:23.496659 4843 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data podName:0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4 nodeName:}" failed. No retries permitted until 2026-03-14 09:55:27.49661973 +0000 UTC m=+2694.809230858 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data") pod "watcher-kuttl-decision-engine-0" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4") : secret "watcher-kuttl-decision-engine-config-data" not found Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.802472 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:23 crc kubenswrapper[4843]: I0314 09:55:23.954115 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerStarted","Data":"cbb9ef6c7f054c2895838b8b669f98eafeae0964c7cf8805305b59c14e768b2e"} Mar 14 09:55:24 crc kubenswrapper[4843]: I0314 09:55:24.262010 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-q467f"] Mar 14 09:55:24 crc kubenswrapper[4843]: I0314 09:55:24.272643 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-q467f"] Mar 14 09:55:24 crc kubenswrapper[4843]: I0314 09:55:24.284199 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-1522-account-create-update-brbfc"] Mar 14 09:55:24 crc kubenswrapper[4843]: I0314 09:55:24.293281 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher1522-account-delete-9vntc"] Mar 14 09:55:24 crc kubenswrapper[4843]: I0314 09:55:24.298530 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-1522-account-create-update-brbfc"] Mar 14 09:55:24 crc kubenswrapper[4843]: I0314 09:55:24.303496 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher1522-account-delete-9vntc"] Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.007734 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerStarted","Data":"b2d7e383568d221919ceb5f7fb6d68e9940cab3453ff30964a3e0734165006b6"} Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.013123 4843 generic.go:334] "Generic (PLEG): container finished" podID="0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" containerID="e4f681c58540489ce55e7cd8f6a8b0ddb618c031e8d78d1c1ac328835e99a228" exitCode=0 Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.013189 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4","Type":"ContainerDied","Data":"e4f681c58540489ce55e7cd8f6a8b0ddb618c031e8d78d1c1ac328835e99a228"} Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.094495 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.230915 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-logs\") pod \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.231011 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-cert-memcached-mtls\") pod \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.231040 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data\") pod \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.231059 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-combined-ca-bundle\") pod \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.231122 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtwdl\" (UniqueName: \"kubernetes.io/projected/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-kube-api-access-qtwdl\") pod \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.231186 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-custom-prometheus-ca\") pod \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\" (UID: \"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4\") " Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.232503 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-logs" (OuterVolumeSpecName: "logs") pod "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.235930 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-kube-api-access-qtwdl" (OuterVolumeSpecName: "kube-api-access-qtwdl") pod "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4"). InnerVolumeSpecName "kube-api-access-qtwdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.254973 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.255439 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.272746 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data" (OuterVolumeSpecName: "config-data") pod "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.300538 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" (UID: "0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.332514 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.332551 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.332560 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.332570 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtwdl\" (UniqueName: \"kubernetes.io/projected/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-kube-api-access-qtwdl\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.332580 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.332588 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.347837 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d271cc6-c547-4e3d-83a3-248d29ace879" path="/var/lib/kubelet/pods/0d271cc6-c547-4e3d-83a3-248d29ace879/volumes" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.348504 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27b43d13-6378-47a9-b3a9-3e9532e47977" path="/var/lib/kubelet/pods/27b43d13-6378-47a9-b3a9-3e9532e47977/volumes" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.349157 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea5456e0-feac-43fe-b412-46cfe1cff4de" path="/var/lib/kubelet/pods/ea5456e0-feac-43fe-b412-46cfe1cff4de/volumes" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.884375 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:25 crc kubenswrapper[4843]: I0314 09:55:25.932119 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.023583 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4","Type":"ContainerDied","Data":"ea5623acbb3ae60cc43c3ef0c979cd0763b24118a241150777f089425ef4e0da"} Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.023643 4843 scope.go:117] "RemoveContainer" containerID="e4f681c58540489ce55e7cd8f6a8b0ddb618c031e8d78d1c1ac328835e99a228" Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.023778 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.028014 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerStarted","Data":"707524ce01e8e00f1e04014c5f3e37688ecd69074d4493e9ffb238c98a088d91"} Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.028190 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerStarted","Data":"fad5bb94f4f1e4b048ce5e1ffad37d22f6906834af47982b8d42c13dfbaf2afa"} Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.053055 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:26 crc kubenswrapper[4843]: I0314 09:55:26.066560 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.376605 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" path="/var/lib/kubelet/pods/0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4/volumes" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.377518 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-db-create-xp85v"] Mar 14 09:55:27 crc kubenswrapper[4843]: E0314 09:55:27.377851 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" containerName="watcher-decision-engine" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.377867 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" containerName="watcher-decision-engine" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.378099 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e98ba2a-93b1-47d4-8e7e-dddb3ca56fc4" containerName="watcher-decision-engine" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.381147 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-xp85v"] Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.381257 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.390196 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-bacc-account-create-update-qzv47"] Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.391345 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.394368 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-db-secret" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.409912 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-bacc-account-create-update-qzv47"] Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.491064 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hrbf\" (UniqueName: \"kubernetes.io/projected/b6f7a856-c96c-4553-84d8-53e419c1d015-kube-api-access-2hrbf\") pod \"watcher-bacc-account-create-update-qzv47\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.492056 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5423a2-a865-4402-90b2-9053439f187b-operator-scripts\") pod \"watcher-db-create-xp85v\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.492452 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r52fx\" (UniqueName: \"kubernetes.io/projected/0c5423a2-a865-4402-90b2-9053439f187b-kube-api-access-r52fx\") pod \"watcher-db-create-xp85v\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.492645 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f7a856-c96c-4553-84d8-53e419c1d015-operator-scripts\") pod \"watcher-bacc-account-create-update-qzv47\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.594681 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f7a856-c96c-4553-84d8-53e419c1d015-operator-scripts\") pod \"watcher-bacc-account-create-update-qzv47\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.594830 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hrbf\" (UniqueName: \"kubernetes.io/projected/b6f7a856-c96c-4553-84d8-53e419c1d015-kube-api-access-2hrbf\") pod \"watcher-bacc-account-create-update-qzv47\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.594939 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5423a2-a865-4402-90b2-9053439f187b-operator-scripts\") pod \"watcher-db-create-xp85v\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.595006 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r52fx\" (UniqueName: \"kubernetes.io/projected/0c5423a2-a865-4402-90b2-9053439f187b-kube-api-access-r52fx\") pod \"watcher-db-create-xp85v\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.595502 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f7a856-c96c-4553-84d8-53e419c1d015-operator-scripts\") pod \"watcher-bacc-account-create-update-qzv47\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.595799 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5423a2-a865-4402-90b2-9053439f187b-operator-scripts\") pod \"watcher-db-create-xp85v\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.610876 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r52fx\" (UniqueName: \"kubernetes.io/projected/0c5423a2-a865-4402-90b2-9053439f187b-kube-api-access-r52fx\") pod \"watcher-db-create-xp85v\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.611532 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hrbf\" (UniqueName: \"kubernetes.io/projected/b6f7a856-c96c-4553-84d8-53e419c1d015-kube-api-access-2hrbf\") pod \"watcher-bacc-account-create-update-qzv47\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.723012 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:27 crc kubenswrapper[4843]: I0314 09:55:27.735051 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:28 crc kubenswrapper[4843]: I0314 09:55:28.048530 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerStarted","Data":"78f3d4945a6094215952ba4551633e30e8e9703f17d741a8718eb310d9545e6f"} Mar 14 09:55:28 crc kubenswrapper[4843]: I0314 09:55:28.048961 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:28 crc kubenswrapper[4843]: I0314 09:55:28.093403 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.854378128 podStartE2EDuration="6.093385043s" podCreationTimestamp="2026-03-14 09:55:22 +0000 UTC" firstStartedPulling="2026-03-14 09:55:23.813895149 +0000 UTC m=+2691.126506297" lastFinishedPulling="2026-03-14 09:55:27.052902084 +0000 UTC m=+2694.365513212" observedRunningTime="2026-03-14 09:55:28.087776586 +0000 UTC m=+2695.400387724" watchObservedRunningTime="2026-03-14 09:55:28.093385043 +0000 UTC m=+2695.405996171" Mar 14 09:55:28 crc kubenswrapper[4843]: I0314 09:55:28.187557 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-bacc-account-create-update-qzv47"] Mar 14 09:55:28 crc kubenswrapper[4843]: I0314 09:55:28.261723 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-db-create-xp85v"] Mar 14 09:55:28 crc kubenswrapper[4843]: W0314 09:55:28.265514 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c5423a2_a865_4402_90b2_9053439f187b.slice/crio-6dd37d79654a381c38e6aa2da5a362c2b99607a578b1893c25abff56ad876592 WatchSource:0}: Error finding container 6dd37d79654a381c38e6aa2da5a362c2b99607a578b1893c25abff56ad876592: Status 404 returned error can't find the container with id 6dd37d79654a381c38e6aa2da5a362c2b99607a578b1893c25abff56ad876592 Mar 14 09:55:29 crc kubenswrapper[4843]: I0314 09:55:29.058189 4843 generic.go:334] "Generic (PLEG): container finished" podID="b6f7a856-c96c-4553-84d8-53e419c1d015" containerID="ccf207a3097e578e80bde5360a5059a4945a7bd296718a419144f6085e14980f" exitCode=0 Mar 14 09:55:29 crc kubenswrapper[4843]: I0314 09:55:29.058286 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" event={"ID":"b6f7a856-c96c-4553-84d8-53e419c1d015","Type":"ContainerDied","Data":"ccf207a3097e578e80bde5360a5059a4945a7bd296718a419144f6085e14980f"} Mar 14 09:55:29 crc kubenswrapper[4843]: I0314 09:55:29.058574 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" event={"ID":"b6f7a856-c96c-4553-84d8-53e419c1d015","Type":"ContainerStarted","Data":"ed894d7d342a9eaf82e33a147d2334b223bc4481c002551b7b7869267d285eb9"} Mar 14 09:55:29 crc kubenswrapper[4843]: I0314 09:55:29.060226 4843 generic.go:334] "Generic (PLEG): container finished" podID="0c5423a2-a865-4402-90b2-9053439f187b" containerID="0fc03499f13ec2dd1367f7d421a08174ac9963e0c5df06c41a130cc4f31ed08d" exitCode=0 Mar 14 09:55:29 crc kubenswrapper[4843]: I0314 09:55:29.060294 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-xp85v" event={"ID":"0c5423a2-a865-4402-90b2-9053439f187b","Type":"ContainerDied","Data":"0fc03499f13ec2dd1367f7d421a08174ac9963e0c5df06c41a130cc4f31ed08d"} Mar 14 09:55:29 crc kubenswrapper[4843]: I0314 09:55:29.060323 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-xp85v" event={"ID":"0c5423a2-a865-4402-90b2-9053439f187b","Type":"ContainerStarted","Data":"6dd37d79654a381c38e6aa2da5a362c2b99607a578b1893c25abff56ad876592"} Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.103034 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j8b6n"] Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.103249 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j8b6n" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="registry-server" containerID="cri-o://473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96" gracePeriod=2 Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.646948 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.653928 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.661009 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755065 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nskfb\" (UniqueName: \"kubernetes.io/projected/4e0036f2-535c-47e5-b145-83c38ea91d71-kube-api-access-nskfb\") pod \"4e0036f2-535c-47e5-b145-83c38ea91d71\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755163 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-catalog-content\") pod \"4e0036f2-535c-47e5-b145-83c38ea91d71\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755210 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r52fx\" (UniqueName: \"kubernetes.io/projected/0c5423a2-a865-4402-90b2-9053439f187b-kube-api-access-r52fx\") pod \"0c5423a2-a865-4402-90b2-9053439f187b\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755242 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-utilities\") pod \"4e0036f2-535c-47e5-b145-83c38ea91d71\" (UID: \"4e0036f2-535c-47e5-b145-83c38ea91d71\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755294 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f7a856-c96c-4553-84d8-53e419c1d015-operator-scripts\") pod \"b6f7a856-c96c-4553-84d8-53e419c1d015\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755363 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5423a2-a865-4402-90b2-9053439f187b-operator-scripts\") pod \"0c5423a2-a865-4402-90b2-9053439f187b\" (UID: \"0c5423a2-a865-4402-90b2-9053439f187b\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755436 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hrbf\" (UniqueName: \"kubernetes.io/projected/b6f7a856-c96c-4553-84d8-53e419c1d015-kube-api-access-2hrbf\") pod \"b6f7a856-c96c-4553-84d8-53e419c1d015\" (UID: \"b6f7a856-c96c-4553-84d8-53e419c1d015\") " Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.755989 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c5423a2-a865-4402-90b2-9053439f187b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c5423a2-a865-4402-90b2-9053439f187b" (UID: "0c5423a2-a865-4402-90b2-9053439f187b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.756084 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f7a856-c96c-4553-84d8-53e419c1d015-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6f7a856-c96c-4553-84d8-53e419c1d015" (UID: "b6f7a856-c96c-4553-84d8-53e419c1d015"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.756504 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-utilities" (OuterVolumeSpecName: "utilities") pod "4e0036f2-535c-47e5-b145-83c38ea91d71" (UID: "4e0036f2-535c-47e5-b145-83c38ea91d71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.760720 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5423a2-a865-4402-90b2-9053439f187b-kube-api-access-r52fx" (OuterVolumeSpecName: "kube-api-access-r52fx") pod "0c5423a2-a865-4402-90b2-9053439f187b" (UID: "0c5423a2-a865-4402-90b2-9053439f187b"). InnerVolumeSpecName "kube-api-access-r52fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.760918 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f7a856-c96c-4553-84d8-53e419c1d015-kube-api-access-2hrbf" (OuterVolumeSpecName: "kube-api-access-2hrbf") pod "b6f7a856-c96c-4553-84d8-53e419c1d015" (UID: "b6f7a856-c96c-4553-84d8-53e419c1d015"). InnerVolumeSpecName "kube-api-access-2hrbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.760917 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e0036f2-535c-47e5-b145-83c38ea91d71-kube-api-access-nskfb" (OuterVolumeSpecName: "kube-api-access-nskfb") pod "4e0036f2-535c-47e5-b145-83c38ea91d71" (UID: "4e0036f2-535c-47e5-b145-83c38ea91d71"). InnerVolumeSpecName "kube-api-access-nskfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.865554 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.865597 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f7a856-c96c-4553-84d8-53e419c1d015-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.865609 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c5423a2-a865-4402-90b2-9053439f187b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.865619 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hrbf\" (UniqueName: \"kubernetes.io/projected/b6f7a856-c96c-4553-84d8-53e419c1d015-kube-api-access-2hrbf\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.865629 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nskfb\" (UniqueName: \"kubernetes.io/projected/4e0036f2-535c-47e5-b145-83c38ea91d71-kube-api-access-nskfb\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.865638 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r52fx\" (UniqueName: \"kubernetes.io/projected/0c5423a2-a865-4402-90b2-9053439f187b-kube-api-access-r52fx\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.905463 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e0036f2-535c-47e5-b145-83c38ea91d71" (UID: "4e0036f2-535c-47e5-b145-83c38ea91d71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:30 crc kubenswrapper[4843]: I0314 09:55:30.967663 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0036f2-535c-47e5-b145-83c38ea91d71-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.079334 4843 generic.go:334] "Generic (PLEG): container finished" podID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerID="473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96" exitCode=0 Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.079427 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerDied","Data":"473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96"} Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.079790 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8b6n" event={"ID":"4e0036f2-535c-47e5-b145-83c38ea91d71","Type":"ContainerDied","Data":"5b3bde5b6a63c969b1462d69f533b4251aa5f69e936390c85ceb01502d290f8a"} Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.079810 4843 scope.go:117] "RemoveContainer" containerID="473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.079465 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8b6n" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.096148 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-db-create-xp85v" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.096177 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-db-create-xp85v" event={"ID":"0c5423a2-a865-4402-90b2-9053439f187b","Type":"ContainerDied","Data":"6dd37d79654a381c38e6aa2da5a362c2b99607a578b1893c25abff56ad876592"} Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.096712 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dd37d79654a381c38e6aa2da5a362c2b99607a578b1893c25abff56ad876592" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.103623 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" event={"ID":"b6f7a856-c96c-4553-84d8-53e419c1d015","Type":"ContainerDied","Data":"ed894d7d342a9eaf82e33a147d2334b223bc4481c002551b7b7869267d285eb9"} Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.103655 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed894d7d342a9eaf82e33a147d2334b223bc4481c002551b7b7869267d285eb9" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.103704 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-bacc-account-create-update-qzv47" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.116555 4843 scope.go:117] "RemoveContainer" containerID="a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.122316 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j8b6n"] Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.128937 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j8b6n"] Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.146573 4843 scope.go:117] "RemoveContainer" containerID="da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.184963 4843 scope.go:117] "RemoveContainer" containerID="473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96" Mar 14 09:55:31 crc kubenswrapper[4843]: E0314 09:55:31.190007 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96\": container with ID starting with 473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96 not found: ID does not exist" containerID="473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.190080 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96"} err="failed to get container status \"473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96\": rpc error: code = NotFound desc = could not find container \"473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96\": container with ID starting with 473103e0062ebe4ed7ee7e694a11a93972b188f7b8e67f2b4ffb333974388b96 not found: ID does not exist" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.190121 4843 scope.go:117] "RemoveContainer" containerID="a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f" Mar 14 09:55:31 crc kubenswrapper[4843]: E0314 09:55:31.191108 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f\": container with ID starting with a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f not found: ID does not exist" containerID="a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.191197 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f"} err="failed to get container status \"a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f\": rpc error: code = NotFound desc = could not find container \"a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f\": container with ID starting with a4b2f9653ad71c6de36b03be04e64b014ba6c452547619cb372be5ba54e3476f not found: ID does not exist" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.191307 4843 scope.go:117] "RemoveContainer" containerID="da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109" Mar 14 09:55:31 crc kubenswrapper[4843]: E0314 09:55:31.191823 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109\": container with ID starting with da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109 not found: ID does not exist" containerID="da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.191871 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109"} err="failed to get container status \"da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109\": rpc error: code = NotFound desc = could not find container \"da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109\": container with ID starting with da3b2e0ff4d38cd90228726adad187ab0c90b460b0c1706d63181a6cb1fac109 not found: ID does not exist" Mar 14 09:55:31 crc kubenswrapper[4843]: E0314 09:55:31.220130 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e0036f2_535c_47e5_b145_83c38ea91d71.slice/crio-5b3bde5b6a63c969b1462d69f533b4251aa5f69e936390c85ceb01502d290f8a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f7a856_c96c_4553_84d8_53e419c1d015.slice/crio-ed894d7d342a9eaf82e33a147d2334b223bc4481c002551b7b7869267d285eb9\": RecentStats: unable to find data in memory cache]" Mar 14 09:55:31 crc kubenswrapper[4843]: I0314 09:55:31.352858 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" path="/var/lib/kubelet/pods/4e0036f2-535c-47e5-b145-83c38ea91d71/volumes" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.593612 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n"] Mar 14 09:55:32 crc kubenswrapper[4843]: E0314 09:55:32.594055 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="extract-content" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594089 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="extract-content" Mar 14 09:55:32 crc kubenswrapper[4843]: E0314 09:55:32.594105 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5423a2-a865-4402-90b2-9053439f187b" containerName="mariadb-database-create" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594128 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5423a2-a865-4402-90b2-9053439f187b" containerName="mariadb-database-create" Mar 14 09:55:32 crc kubenswrapper[4843]: E0314 09:55:32.594147 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="registry-server" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594153 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="registry-server" Mar 14 09:55:32 crc kubenswrapper[4843]: E0314 09:55:32.594159 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f7a856-c96c-4553-84d8-53e419c1d015" containerName="mariadb-account-create-update" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594165 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f7a856-c96c-4553-84d8-53e419c1d015" containerName="mariadb-account-create-update" Mar 14 09:55:32 crc kubenswrapper[4843]: E0314 09:55:32.594176 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="extract-utilities" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594181 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="extract-utilities" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594346 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5423a2-a865-4402-90b2-9053439f187b" containerName="mariadb-database-create" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594369 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f7a856-c96c-4553-84d8-53e419c1d015" containerName="mariadb-account-create-update" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594384 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e0036f2-535c-47e5-b145-83c38ea91d71" containerName="registry-server" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.594931 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.597103 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-config-data" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.597472 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-jqdkg" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.601077 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n"] Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.744851 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-config-data\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.744895 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.744937 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.744987 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhhc5\" (UniqueName: \"kubernetes.io/projected/4083133e-f6cd-4ec6-8a41-61c440fa81c2-kube-api-access-fhhc5\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.846158 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-config-data\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.846205 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.846242 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.846286 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhhc5\" (UniqueName: \"kubernetes.io/projected/4083133e-f6cd-4ec6-8a41-61c440fa81c2-kube-api-access-fhhc5\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.853881 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-db-sync-config-data\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.854180 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-config-data\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.856873 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-combined-ca-bundle\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.873047 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhhc5\" (UniqueName: \"kubernetes.io/projected/4083133e-f6cd-4ec6-8a41-61c440fa81c2-kube-api-access-fhhc5\") pod \"watcher-kuttl-db-sync-bkc7n\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:32 crc kubenswrapper[4843]: I0314 09:55:32.953854 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:33 crc kubenswrapper[4843]: I0314 09:55:33.398417 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n"] Mar 14 09:55:34 crc kubenswrapper[4843]: I0314 09:55:34.130803 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" event={"ID":"4083133e-f6cd-4ec6-8a41-61c440fa81c2","Type":"ContainerStarted","Data":"3d8f7c84a9f2927c286166a5bf0d0ea9bf0bc5825f41f68d315f57ccad05294e"} Mar 14 09:55:34 crc kubenswrapper[4843]: I0314 09:55:34.131129 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" event={"ID":"4083133e-f6cd-4ec6-8a41-61c440fa81c2","Type":"ContainerStarted","Data":"bebefe57d51b1e32903b0826a1fb0f11164ad1b369d47b34da560058a7e53fe5"} Mar 14 09:55:34 crc kubenswrapper[4843]: I0314 09:55:34.151299 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" podStartSLOduration=2.151266136 podStartE2EDuration="2.151266136s" podCreationTimestamp="2026-03-14 09:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:55:34.149565455 +0000 UTC m=+2701.462176593" watchObservedRunningTime="2026-03-14 09:55:34.151266136 +0000 UTC m=+2701.463877264" Mar 14 09:55:36 crc kubenswrapper[4843]: I0314 09:55:36.146888 4843 generic.go:334] "Generic (PLEG): container finished" podID="4083133e-f6cd-4ec6-8a41-61c440fa81c2" containerID="3d8f7c84a9f2927c286166a5bf0d0ea9bf0bc5825f41f68d315f57ccad05294e" exitCode=0 Mar 14 09:55:36 crc kubenswrapper[4843]: I0314 09:55:36.146968 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" event={"ID":"4083133e-f6cd-4ec6-8a41-61c440fa81c2","Type":"ContainerDied","Data":"3d8f7c84a9f2927c286166a5bf0d0ea9bf0bc5825f41f68d315f57ccad05294e"} Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.521721 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.625328 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-combined-ca-bundle\") pod \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.625436 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-config-data\") pod \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.625464 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-db-sync-config-data\") pod \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.625573 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhhc5\" (UniqueName: \"kubernetes.io/projected/4083133e-f6cd-4ec6-8a41-61c440fa81c2-kube-api-access-fhhc5\") pod \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\" (UID: \"4083133e-f6cd-4ec6-8a41-61c440fa81c2\") " Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.630435 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4083133e-f6cd-4ec6-8a41-61c440fa81c2-kube-api-access-fhhc5" (OuterVolumeSpecName: "kube-api-access-fhhc5") pod "4083133e-f6cd-4ec6-8a41-61c440fa81c2" (UID: "4083133e-f6cd-4ec6-8a41-61c440fa81c2"). InnerVolumeSpecName "kube-api-access-fhhc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.632393 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4083133e-f6cd-4ec6-8a41-61c440fa81c2" (UID: "4083133e-f6cd-4ec6-8a41-61c440fa81c2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.651289 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4083133e-f6cd-4ec6-8a41-61c440fa81c2" (UID: "4083133e-f6cd-4ec6-8a41-61c440fa81c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.691982 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-config-data" (OuterVolumeSpecName: "config-data") pod "4083133e-f6cd-4ec6-8a41-61c440fa81c2" (UID: "4083133e-f6cd-4ec6-8a41-61c440fa81c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.727962 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.728004 4843 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.728021 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhhc5\" (UniqueName: \"kubernetes.io/projected/4083133e-f6cd-4ec6-8a41-61c440fa81c2-kube-api-access-fhhc5\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:37 crc kubenswrapper[4843]: I0314 09:55:37.728035 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4083133e-f6cd-4ec6-8a41-61c440fa81c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.165996 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" event={"ID":"4083133e-f6cd-4ec6-8a41-61c440fa81c2","Type":"ContainerDied","Data":"bebefe57d51b1e32903b0826a1fb0f11164ad1b369d47b34da560058a7e53fe5"} Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.166040 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bebefe57d51b1e32903b0826a1fb0f11164ad1b369d47b34da560058a7e53fe5" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.166098 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.441317 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:38 crc kubenswrapper[4843]: E0314 09:55:38.441809 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4083133e-f6cd-4ec6-8a41-61c440fa81c2" containerName="watcher-kuttl-db-sync" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.441823 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4083133e-f6cd-4ec6-8a41-61c440fa81c2" containerName="watcher-kuttl-db-sync" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.441979 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4083133e-f6cd-4ec6-8a41-61c440fa81c2" containerName="watcher-kuttl-db-sync" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.442506 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.444733 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-watcher-kuttl-dockercfg-jqdkg" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.444956 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-decision-engine-config-data" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.462387 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.477715 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.478692 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.483007 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-applier-config-data" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.540784 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.540843 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq9pk\" (UniqueName: \"kubernetes.io/projected/0fb16236-e0bb-4b07-ae82-f10c19784b25-kube-api-access-bq9pk\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.540876 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb16236-e0bb-4b07-ae82-f10c19784b25-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.540903 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.540948 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.541030 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.541893 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.576000 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.577702 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.583005 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"watcher-kuttl-api-config-data" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.598912 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.600809 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.620384 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.632617 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643704 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643749 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2ce34bd-848a-451d-b698-cb9e208aec88-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643806 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643826 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p55m4\" (UniqueName: \"kubernetes.io/projected/b2ce34bd-848a-451d-b698-cb9e208aec88-kube-api-access-p55m4\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643868 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643889 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643906 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq9pk\" (UniqueName: \"kubernetes.io/projected/0fb16236-e0bb-4b07-ae82-f10c19784b25-kube-api-access-bq9pk\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643924 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb16236-e0bb-4b07-ae82-f10c19784b25-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643946 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643963 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.643997 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.645031 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb16236-e0bb-4b07-ae82-f10c19784b25-logs\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.647611 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-cert-memcached-mtls\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.652542 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-custom-prometheus-ca\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.652753 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-config-data\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.664616 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq9pk\" (UniqueName: \"kubernetes.io/projected/0fb16236-e0bb-4b07-ae82-f10c19784b25-kube-api-access-bq9pk\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.665544 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-combined-ca-bundle\") pod \"watcher-kuttl-decision-engine-0\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745081 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745145 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745179 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lb68\" (UniqueName: \"kubernetes.io/projected/94a9e1eb-8774-41d4-98fa-ea1112f21740-kube-api-access-5lb68\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745213 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745251 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745294 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94a9e1eb-8774-41d4-98fa-ea1112f21740-logs\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745315 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745351 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745378 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745403 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2ce34bd-848a-451d-b698-cb9e208aec88-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745428 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745451 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745472 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745495 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.745524 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p59b\" (UniqueName: \"kubernetes.io/projected/843aaf0d-9d6e-486d-91b2-020cadb3b194-kube-api-access-7p59b\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.746052 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2ce34bd-848a-451d-b698-cb9e208aec88-logs\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.746180 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p55m4\" (UniqueName: \"kubernetes.io/projected/b2ce34bd-848a-451d-b698-cb9e208aec88-kube-api-access-p55m4\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.746496 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/843aaf0d-9d6e-486d-91b2-020cadb3b194-logs\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.748175 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-cert-memcached-mtls\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.748872 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-config-data\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.750665 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-combined-ca-bundle\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.761934 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.763926 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p55m4\" (UniqueName: \"kubernetes.io/projected/b2ce34bd-848a-451d-b698-cb9e208aec88-kube-api-access-p55m4\") pod \"watcher-kuttl-applier-0\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.830913 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849732 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849799 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lb68\" (UniqueName: \"kubernetes.io/projected/94a9e1eb-8774-41d4-98fa-ea1112f21740-kube-api-access-5lb68\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849848 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849876 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849898 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94a9e1eb-8774-41d4-98fa-ea1112f21740-logs\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849933 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849966 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.849986 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.850008 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.850028 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.850059 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p59b\" (UniqueName: \"kubernetes.io/projected/843aaf0d-9d6e-486d-91b2-020cadb3b194-kube-api-access-7p59b\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.850116 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/843aaf0d-9d6e-486d-91b2-020cadb3b194-logs\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.851303 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/843aaf0d-9d6e-486d-91b2-020cadb3b194-logs\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.853306 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94a9e1eb-8774-41d4-98fa-ea1112f21740-logs\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.860563 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-combined-ca-bundle\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.869110 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.869400 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-custom-prometheus-ca\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.869455 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-custom-prometheus-ca\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.875339 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-config-data\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.876222 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-config-data\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.882149 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-cert-memcached-mtls\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.882571 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lb68\" (UniqueName: \"kubernetes.io/projected/94a9e1eb-8774-41d4-98fa-ea1112f21740-kube-api-access-5lb68\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.886965 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-cert-memcached-mtls\") pod \"watcher-kuttl-api-1\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.887720 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p59b\" (UniqueName: \"kubernetes.io/projected/843aaf0d-9d6e-486d-91b2-020cadb3b194-kube-api-access-7p59b\") pod \"watcher-kuttl-api-0\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.913876 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:38 crc kubenswrapper[4843]: I0314 09:55:38.930127 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:39 crc kubenswrapper[4843]: I0314 09:55:39.256617 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:55:39 crc kubenswrapper[4843]: W0314 09:55:39.262610 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94a9e1eb_8774_41d4_98fa_ea1112f21740.slice/crio-2728517a9eb41ffdbc3a3dc492c1552f5cdbb345372d909a560331112f3da9bd WatchSource:0}: Error finding container 2728517a9eb41ffdbc3a3dc492c1552f5cdbb345372d909a560331112f3da9bd: Status 404 returned error can't find the container with id 2728517a9eb41ffdbc3a3dc492c1552f5cdbb345372d909a560331112f3da9bd Mar 14 09:55:39 crc kubenswrapper[4843]: I0314 09:55:39.285574 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:55:39 crc kubenswrapper[4843]: W0314 09:55:39.303897 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fb16236_e0bb_4b07_ae82_f10c19784b25.slice/crio-60e1cd04eda923dbfacdcb061dea421b85de46f713a04ce833e9c5b362bb6ab0 WatchSource:0}: Error finding container 60e1cd04eda923dbfacdcb061dea421b85de46f713a04ce833e9c5b362bb6ab0: Status 404 returned error can't find the container with id 60e1cd04eda923dbfacdcb061dea421b85de46f713a04ce833e9c5b362bb6ab0 Mar 14 09:55:39 crc kubenswrapper[4843]: I0314 09:55:39.382660 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:55:39 crc kubenswrapper[4843]: I0314 09:55:39.570371 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.182237 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"94a9e1eb-8774-41d4-98fa-ea1112f21740","Type":"ContainerStarted","Data":"df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.182532 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"94a9e1eb-8774-41d4-98fa-ea1112f21740","Type":"ContainerStarted","Data":"949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.182545 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"94a9e1eb-8774-41d4-98fa-ea1112f21740","Type":"ContainerStarted","Data":"2728517a9eb41ffdbc3a3dc492c1552f5cdbb345372d909a560331112f3da9bd"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.182559 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.184551 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2ce34bd-848a-451d-b698-cb9e208aec88","Type":"ContainerStarted","Data":"a121047c5cdf4ad35daa78b13fb464c5fa84f92d09ccd66907cdcde562dfa4a7"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.184588 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2ce34bd-848a-451d-b698-cb9e208aec88","Type":"ContainerStarted","Data":"db495bd932d770248f5787cf5eeb8cf85e3d612bee49a5a61c5d795e659b168a"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.187287 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0fb16236-e0bb-4b07-ae82-f10c19784b25","Type":"ContainerStarted","Data":"09034172c7e541ed0e50b75dd991a61b4728706d7abc69c0f4cfb74c97d8b87a"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.187313 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0fb16236-e0bb-4b07-ae82-f10c19784b25","Type":"ContainerStarted","Data":"60e1cd04eda923dbfacdcb061dea421b85de46f713a04ce833e9c5b362bb6ab0"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.189190 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"843aaf0d-9d6e-486d-91b2-020cadb3b194","Type":"ContainerStarted","Data":"7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.189236 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"843aaf0d-9d6e-486d-91b2-020cadb3b194","Type":"ContainerStarted","Data":"01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.189254 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"843aaf0d-9d6e-486d-91b2-020cadb3b194","Type":"ContainerStarted","Data":"c6b74c8bf8741ef504b611dac95dcb5d267359170d0e1ac614fab76048b55dc6"} Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.189523 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.190230 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.1.13:9322/\": dial tcp 10.217.1.13:9322: connect: connection refused" Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.209241 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-1" podStartSLOduration=2.209215281 podStartE2EDuration="2.209215281s" podCreationTimestamp="2026-03-14 09:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:55:40.205119981 +0000 UTC m=+2707.517731109" watchObservedRunningTime="2026-03-14 09:55:40.209215281 +0000 UTC m=+2707.521826409" Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.223150 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podStartSLOduration=2.223132692 podStartE2EDuration="2.223132692s" podCreationTimestamp="2026-03-14 09:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:55:40.22181931 +0000 UTC m=+2707.534430428" watchObservedRunningTime="2026-03-14 09:55:40.223132692 +0000 UTC m=+2707.535743820" Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.245608 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podStartSLOduration=2.245592862 podStartE2EDuration="2.245592862s" podCreationTimestamp="2026-03-14 09:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:55:40.242437764 +0000 UTC m=+2707.555048882" watchObservedRunningTime="2026-03-14 09:55:40.245592862 +0000 UTC m=+2707.558203980" Mar 14 09:55:40 crc kubenswrapper[4843]: I0314 09:55:40.264394 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-0" podStartSLOduration=2.264375292 podStartE2EDuration="2.264375292s" podCreationTimestamp="2026-03-14 09:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:55:40.262601208 +0000 UTC m=+2707.575212336" watchObservedRunningTime="2026-03-14 09:55:40.264375292 +0000 UTC m=+2707.576986420" Mar 14 09:55:42 crc kubenswrapper[4843]: I0314 09:55:42.463904 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:43 crc kubenswrapper[4843]: I0314 09:55:43.411750 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:43 crc kubenswrapper[4843]: I0314 09:55:43.832090 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:43 crc kubenswrapper[4843]: I0314 09:55:43.914793 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:43 crc kubenswrapper[4843]: I0314 09:55:43.930424 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:45 crc kubenswrapper[4843]: I0314 09:55:45.388079 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:55:45 crc kubenswrapper[4843]: I0314 09:55:45.388434 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.763556 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.831723 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.843915 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.872310 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.914388 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.930301 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.930439 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:48 crc kubenswrapper[4843]: I0314 09:55:48.949779 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.273877 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.278935 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.279561 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.299754 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.301108 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.515561 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dqj9j"] Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.519560 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.534637 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dqj9j"] Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.652788 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thnch\" (UniqueName: \"kubernetes.io/projected/45d44dc1-12d7-422e-acbf-1a2f5008237b-kube-api-access-thnch\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.653240 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-utilities\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.653354 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-catalog-content\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.754860 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thnch\" (UniqueName: \"kubernetes.io/projected/45d44dc1-12d7-422e-acbf-1a2f5008237b-kube-api-access-thnch\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.755026 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-utilities\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.755062 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-catalog-content\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.755564 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-utilities\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.755620 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-catalog-content\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.775844 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thnch\" (UniqueName: \"kubernetes.io/projected/45d44dc1-12d7-422e-acbf-1a2f5008237b-kube-api-access-thnch\") pod \"redhat-marketplace-dqj9j\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:49 crc kubenswrapper[4843]: I0314 09:55:49.848055 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.349831 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dqj9j"] Mar 14 09:55:50 crc kubenswrapper[4843]: W0314 09:55:50.362037 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45d44dc1_12d7_422e_acbf_1a2f5008237b.slice/crio-55ed977c5f95a97636006d6db113bbc6fd24dc1a81b0a5624ada19d8f534e2fc WatchSource:0}: Error finding container 55ed977c5f95a97636006d6db113bbc6fd24dc1a81b0a5624ada19d8f534e2fc: Status 404 returned error can't find the container with id 55ed977c5f95a97636006d6db113bbc6fd24dc1a81b0a5624ada19d8f534e2fc Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.812217 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.812911 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-central-agent" containerID="cri-o://b2d7e383568d221919ceb5f7fb6d68e9940cab3453ff30964a3e0734165006b6" gracePeriod=30 Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.813729 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-notification-agent" containerID="cri-o://fad5bb94f4f1e4b048ce5e1ffad37d22f6906834af47982b8d42c13dfbaf2afa" gracePeriod=30 Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.813859 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="proxy-httpd" containerID="cri-o://78f3d4945a6094215952ba4551633e30e8e9703f17d741a8718eb310d9545e6f" gracePeriod=30 Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.813847 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="sg-core" containerID="cri-o://707524ce01e8e00f1e04014c5f3e37688ecd69074d4493e9ffb238c98a088d91" gracePeriod=30 Mar 14 09:55:50 crc kubenswrapper[4843]: I0314 09:55:50.836442 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.7:3000/\": read tcp 10.217.0.2:55022->10.217.1.7:3000: read: connection reset by peer" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303537 4843 generic.go:334] "Generic (PLEG): container finished" podID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerID="78f3d4945a6094215952ba4551633e30e8e9703f17d741a8718eb310d9545e6f" exitCode=0 Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303570 4843 generic.go:334] "Generic (PLEG): container finished" podID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerID="707524ce01e8e00f1e04014c5f3e37688ecd69074d4493e9ffb238c98a088d91" exitCode=2 Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303579 4843 generic.go:334] "Generic (PLEG): container finished" podID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerID="fad5bb94f4f1e4b048ce5e1ffad37d22f6906834af47982b8d42c13dfbaf2afa" exitCode=0 Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303589 4843 generic.go:334] "Generic (PLEG): container finished" podID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerID="b2d7e383568d221919ceb5f7fb6d68e9940cab3453ff30964a3e0734165006b6" exitCode=0 Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303623 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerDied","Data":"78f3d4945a6094215952ba4551633e30e8e9703f17d741a8718eb310d9545e6f"} Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303680 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerDied","Data":"707524ce01e8e00f1e04014c5f3e37688ecd69074d4493e9ffb238c98a088d91"} Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303695 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerDied","Data":"fad5bb94f4f1e4b048ce5e1ffad37d22f6906834af47982b8d42c13dfbaf2afa"} Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.303706 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerDied","Data":"b2d7e383568d221919ceb5f7fb6d68e9940cab3453ff30964a3e0734165006b6"} Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.305884 4843 generic.go:334] "Generic (PLEG): container finished" podID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerID="f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e" exitCode=0 Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.306781 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerDied","Data":"f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e"} Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.306821 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerStarted","Data":"55ed977c5f95a97636006d6db113bbc6fd24dc1a81b0a5624ada19d8f534e2fc"} Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.307962 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.769666 4843 scope.go:117] "RemoveContainer" containerID="14dfef585e74df4bd8cb39675afb6b50972f6c6d844225d3abbf6e5a6d2cbb26" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.829064 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.830986 4843 scope.go:117] "RemoveContainer" containerID="b0814bae3cb1837930d651cccd41bb5520013bb18641d5a98b8fcb4fdcd647b4" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.850461 4843 scope.go:117] "RemoveContainer" containerID="ad2888b2c4d2a3bfb91ed2f3be3d9eca1437462b443ff21fb53a06decb96c392" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.879859 4843 scope.go:117] "RemoveContainer" containerID="bd02424abe7d682a36e22ac815021887580e3be4a905e7e179e343c7815d4a88" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.901139 4843 scope.go:117] "RemoveContainer" containerID="cb97cad51ce45cd869003032d25e4fd917ed6e7ba21bff54560f41ed7b1ce607" Mar 14 09:55:51 crc kubenswrapper[4843]: I0314 09:55:51.949836 4843 scope.go:117] "RemoveContainer" containerID="b63f9b1db597dc2dad973419315a3bc0625aeeb62ed1e2d467bf0ac2f49f913e" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016488 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-run-httpd\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016615 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-sg-core-conf-yaml\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016663 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q5jg\" (UniqueName: \"kubernetes.io/projected/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-kube-api-access-7q5jg\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016691 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-config-data\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016738 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-log-httpd\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016779 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-combined-ca-bundle\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016924 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-ceilometer-tls-certs\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.016971 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-scripts\") pod \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\" (UID: \"4860a55c-8ce8-470b-962e-9ff19f7d8d7b\") " Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.019892 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.022165 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.027432 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-kube-api-access-7q5jg" (OuterVolumeSpecName: "kube-api-access-7q5jg") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "kube-api-access-7q5jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.032491 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-scripts" (OuterVolumeSpecName: "scripts") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.057383 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.072888 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.096455 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119459 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119560 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q5jg\" (UniqueName: \"kubernetes.io/projected/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-kube-api-access-7q5jg\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119576 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119589 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119602 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119613 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.119623 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.122682 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-config-data" (OuterVolumeSpecName: "config-data") pod "4860a55c-8ce8-470b-962e-9ff19f7d8d7b" (UID: "4860a55c-8ce8-470b-962e-9ff19f7d8d7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.221321 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4860a55c-8ce8-470b-962e-9ff19f7d8d7b-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.316880 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.316867 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"4860a55c-8ce8-470b-962e-9ff19f7d8d7b","Type":"ContainerDied","Data":"cbb9ef6c7f054c2895838b8b669f98eafeae0964c7cf8805305b59c14e768b2e"} Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.317068 4843 scope.go:117] "RemoveContainer" containerID="78f3d4945a6094215952ba4551633e30e8e9703f17d741a8718eb310d9545e6f" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.319636 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerStarted","Data":"0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3"} Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.346720 4843 scope.go:117] "RemoveContainer" containerID="707524ce01e8e00f1e04014c5f3e37688ecd69074d4493e9ffb238c98a088d91" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.362205 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.383206 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.392712 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:52 crc kubenswrapper[4843]: E0314 09:55:52.393589 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-notification-agent" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393608 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-notification-agent" Mar 14 09:55:52 crc kubenswrapper[4843]: E0314 09:55:52.393669 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="proxy-httpd" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393677 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="proxy-httpd" Mar 14 09:55:52 crc kubenswrapper[4843]: E0314 09:55:52.393688 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="sg-core" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393695 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="sg-core" Mar 14 09:55:52 crc kubenswrapper[4843]: E0314 09:55:52.393705 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-central-agent" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393711 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-central-agent" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393855 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-notification-agent" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393865 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="sg-core" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393879 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="ceilometer-central-agent" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.393888 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" containerName="proxy-httpd" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.395256 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.399460 4843 scope.go:117] "RemoveContainer" containerID="fad5bb94f4f1e4b048ce5e1ffad37d22f6906834af47982b8d42c13dfbaf2afa" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.403651 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.403771 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.403877 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.409198 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.470568 4843 scope.go:117] "RemoveContainer" containerID="b2d7e383568d221919ceb5f7fb6d68e9940cab3453ff30964a3e0734165006b6" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533540 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-scripts\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533627 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrkm\" (UniqueName: \"kubernetes.io/projected/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-kube-api-access-ctrkm\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533669 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-log-httpd\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533719 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533743 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533766 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-run-httpd\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533797 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-config-data\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.533841 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635583 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctrkm\" (UniqueName: \"kubernetes.io/projected/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-kube-api-access-ctrkm\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635658 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-log-httpd\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635707 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635731 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635757 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-run-httpd\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635794 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-config-data\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635839 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.635877 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-scripts\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.636903 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-run-httpd\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.637183 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-log-httpd\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.641830 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.642096 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-scripts\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.643457 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.644966 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.645689 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-config-data\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.655120 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctrkm\" (UniqueName: \"kubernetes.io/projected/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-kube-api-access-ctrkm\") pod \"ceilometer-0\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:52 crc kubenswrapper[4843]: I0314 09:55:52.772084 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:53 crc kubenswrapper[4843]: W0314 09:55:53.214886 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00e2f129_cb1f_48d7_bb47_603ec1bbdec7.slice/crio-33cfed407eae5920a287b73b7758495fd772838906f797dc7b04491e2994bb3a WatchSource:0}: Error finding container 33cfed407eae5920a287b73b7758495fd772838906f797dc7b04491e2994bb3a: Status 404 returned error can't find the container with id 33cfed407eae5920a287b73b7758495fd772838906f797dc7b04491e2994bb3a Mar 14 09:55:53 crc kubenswrapper[4843]: I0314 09:55:53.221968 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:55:53 crc kubenswrapper[4843]: I0314 09:55:53.327455 4843 generic.go:334] "Generic (PLEG): container finished" podID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerID="0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3" exitCode=0 Mar 14 09:55:53 crc kubenswrapper[4843]: I0314 09:55:53.327508 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerDied","Data":"0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3"} Mar 14 09:55:53 crc kubenswrapper[4843]: I0314 09:55:53.330592 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerStarted","Data":"33cfed407eae5920a287b73b7758495fd772838906f797dc7b04491e2994bb3a"} Mar 14 09:55:53 crc kubenswrapper[4843]: I0314 09:55:53.349986 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4860a55c-8ce8-470b-962e-9ff19f7d8d7b" path="/var/lib/kubelet/pods/4860a55c-8ce8-470b-962e-9ff19f7d8d7b/volumes" Mar 14 09:55:54 crc kubenswrapper[4843]: I0314 09:55:54.345647 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerStarted","Data":"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472"} Mar 14 09:55:54 crc kubenswrapper[4843]: I0314 09:55:54.348390 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerStarted","Data":"a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9"} Mar 14 09:55:54 crc kubenswrapper[4843]: I0314 09:55:54.374019 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dqj9j" podStartSLOduration=2.829173296 podStartE2EDuration="5.373993233s" podCreationTimestamp="2026-03-14 09:55:49 +0000 UTC" firstStartedPulling="2026-03-14 09:55:51.307760948 +0000 UTC m=+2718.620372076" lastFinishedPulling="2026-03-14 09:55:53.852580885 +0000 UTC m=+2721.165192013" observedRunningTime="2026-03-14 09:55:54.366791637 +0000 UTC m=+2721.679402765" watchObservedRunningTime="2026-03-14 09:55:54.373993233 +0000 UTC m=+2721.686604361" Mar 14 09:55:55 crc kubenswrapper[4843]: I0314 09:55:55.362369 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerStarted","Data":"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9"} Mar 14 09:55:56 crc kubenswrapper[4843]: I0314 09:55:56.375136 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerStarted","Data":"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c"} Mar 14 09:55:58 crc kubenswrapper[4843]: I0314 09:55:58.392794 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerStarted","Data":"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50"} Mar 14 09:55:58 crc kubenswrapper[4843]: I0314 09:55:58.393580 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:55:58 crc kubenswrapper[4843]: I0314 09:55:58.420082 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.459995209 podStartE2EDuration="6.420062552s" podCreationTimestamp="2026-03-14 09:55:52 +0000 UTC" firstStartedPulling="2026-03-14 09:55:53.217248637 +0000 UTC m=+2720.529859765" lastFinishedPulling="2026-03-14 09:55:57.17731598 +0000 UTC m=+2724.489927108" observedRunningTime="2026-03-14 09:55:58.413554513 +0000 UTC m=+2725.726165651" watchObservedRunningTime="2026-03-14 09:55:58.420062552 +0000 UTC m=+2725.732673700" Mar 14 09:55:59 crc kubenswrapper[4843]: I0314 09:55:59.849068 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:59 crc kubenswrapper[4843]: I0314 09:55:59.849385 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:55:59 crc kubenswrapper[4843]: I0314 09:55:59.927578 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.139091 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558036-9tltz"] Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.140612 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.142914 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.143302 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.147829 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558036-9tltz"] Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.158243 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.202661 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmhcb\" (UniqueName: \"kubernetes.io/projected/367c08b3-ceb2-4787-841b-c8bb6d3ce971-kube-api-access-tmhcb\") pod \"auto-csr-approver-29558036-9tltz\" (UID: \"367c08b3-ceb2-4787-841b-c8bb6d3ce971\") " pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.304053 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmhcb\" (UniqueName: \"kubernetes.io/projected/367c08b3-ceb2-4787-841b-c8bb6d3ce971-kube-api-access-tmhcb\") pod \"auto-csr-approver-29558036-9tltz\" (UID: \"367c08b3-ceb2-4787-841b-c8bb6d3ce971\") " pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.326202 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmhcb\" (UniqueName: \"kubernetes.io/projected/367c08b3-ceb2-4787-841b-c8bb6d3ce971-kube-api-access-tmhcb\") pod \"auto-csr-approver-29558036-9tltz\" (UID: \"367c08b3-ceb2-4787-841b-c8bb6d3ce971\") " pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.459954 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.466838 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:00 crc kubenswrapper[4843]: I0314 09:56:00.949067 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558036-9tltz"] Mar 14 09:56:00 crc kubenswrapper[4843]: W0314 09:56:00.961048 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod367c08b3_ceb2_4787_841b_c8bb6d3ce971.slice/crio-d1b1644f1991b363a4fafeb2de745904c0856564ef87bd30e3185a21a2dd5fb1 WatchSource:0}: Error finding container d1b1644f1991b363a4fafeb2de745904c0856564ef87bd30e3185a21a2dd5fb1: Status 404 returned error can't find the container with id d1b1644f1991b363a4fafeb2de745904c0856564ef87bd30e3185a21a2dd5fb1 Mar 14 09:56:01 crc kubenswrapper[4843]: I0314 09:56:01.418142 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558036-9tltz" event={"ID":"367c08b3-ceb2-4787-841b-c8bb6d3ce971","Type":"ContainerStarted","Data":"d1b1644f1991b363a4fafeb2de745904c0856564ef87bd30e3185a21a2dd5fb1"} Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.428178 4843 generic.go:334] "Generic (PLEG): container finished" podID="367c08b3-ceb2-4787-841b-c8bb6d3ce971" containerID="3f37680f9b8e4cdee591aea42a93e501cdf46ecde5506a6f974159bc264de966" exitCode=0 Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.428243 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558036-9tltz" event={"ID":"367c08b3-ceb2-4787-841b-c8bb6d3ce971","Type":"ContainerDied","Data":"3f37680f9b8e4cdee591aea42a93e501cdf46ecde5506a6f974159bc264de966"} Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.504504 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.506464 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.523067 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.543350 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-custom-prometheus-ca\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.543410 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m9gk\" (UniqueName: \"kubernetes.io/projected/28affc1b-492c-4671-b111-d2d1d5393cc6-kube-api-access-7m9gk\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.543475 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-combined-ca-bundle\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.543499 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28affc1b-492c-4671-b111-d2d1d5393cc6-logs\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.543554 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-config-data\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.543595 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-cert-memcached-mtls\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.645394 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m9gk\" (UniqueName: \"kubernetes.io/projected/28affc1b-492c-4671-b111-d2d1d5393cc6-kube-api-access-7m9gk\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.645478 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-combined-ca-bundle\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.645503 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28affc1b-492c-4671-b111-d2d1d5393cc6-logs\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.645548 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-config-data\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.645580 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-cert-memcached-mtls\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.645634 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-custom-prometheus-ca\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.646844 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28affc1b-492c-4671-b111-d2d1d5393cc6-logs\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.651829 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-config-data\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.652302 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-custom-prometheus-ca\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.652311 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-cert-memcached-mtls\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.653073 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-combined-ca-bundle\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.673786 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m9gk\" (UniqueName: \"kubernetes.io/projected/28affc1b-492c-4671-b111-d2d1d5393cc6-kube-api-access-7m9gk\") pod \"watcher-kuttl-api-2\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:02 crc kubenswrapper[4843]: I0314 09:56:02.829945 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:03 crc kubenswrapper[4843]: W0314 09:56:03.439440 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28affc1b_492c_4671_b111_d2d1d5393cc6.slice/crio-459b0f030dac43b6b6e9a03039e3ce3ac85d62e623226079874e06b794a4d1cd WatchSource:0}: Error finding container 459b0f030dac43b6b6e9a03039e3ce3ac85d62e623226079874e06b794a4d1cd: Status 404 returned error can't find the container with id 459b0f030dac43b6b6e9a03039e3ce3ac85d62e623226079874e06b794a4d1cd Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.455620 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.505260 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dqj9j"] Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.505667 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dqj9j" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="registry-server" containerID="cri-o://a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9" gracePeriod=2 Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.819717 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.866740 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmhcb\" (UniqueName: \"kubernetes.io/projected/367c08b3-ceb2-4787-841b-c8bb6d3ce971-kube-api-access-tmhcb\") pod \"367c08b3-ceb2-4787-841b-c8bb6d3ce971\" (UID: \"367c08b3-ceb2-4787-841b-c8bb6d3ce971\") " Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.876696 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/367c08b3-ceb2-4787-841b-c8bb6d3ce971-kube-api-access-tmhcb" (OuterVolumeSpecName: "kube-api-access-tmhcb") pod "367c08b3-ceb2-4787-841b-c8bb6d3ce971" (UID: "367c08b3-ceb2-4787-841b-c8bb6d3ce971"). InnerVolumeSpecName "kube-api-access-tmhcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:03 crc kubenswrapper[4843]: I0314 09:56:03.969060 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmhcb\" (UniqueName: \"kubernetes.io/projected/367c08b3-ceb2-4787-841b-c8bb6d3ce971-kube-api-access-tmhcb\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.094025 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.272701 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-utilities\") pod \"45d44dc1-12d7-422e-acbf-1a2f5008237b\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.272898 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-catalog-content\") pod \"45d44dc1-12d7-422e-acbf-1a2f5008237b\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.273901 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-utilities" (OuterVolumeSpecName: "utilities") pod "45d44dc1-12d7-422e-acbf-1a2f5008237b" (UID: "45d44dc1-12d7-422e-acbf-1a2f5008237b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.277412 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thnch\" (UniqueName: \"kubernetes.io/projected/45d44dc1-12d7-422e-acbf-1a2f5008237b-kube-api-access-thnch\") pod \"45d44dc1-12d7-422e-acbf-1a2f5008237b\" (UID: \"45d44dc1-12d7-422e-acbf-1a2f5008237b\") " Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.278097 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.291481 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d44dc1-12d7-422e-acbf-1a2f5008237b-kube-api-access-thnch" (OuterVolumeSpecName: "kube-api-access-thnch") pod "45d44dc1-12d7-422e-acbf-1a2f5008237b" (UID: "45d44dc1-12d7-422e-acbf-1a2f5008237b"). InnerVolumeSpecName "kube-api-access-thnch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.310614 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45d44dc1-12d7-422e-acbf-1a2f5008237b" (UID: "45d44dc1-12d7-422e-acbf-1a2f5008237b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.379537 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45d44dc1-12d7-422e-acbf-1a2f5008237b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.380534 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thnch\" (UniqueName: \"kubernetes.io/projected/45d44dc1-12d7-422e-acbf-1a2f5008237b-kube-api-access-thnch\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.459774 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"28affc1b-492c-4671-b111-d2d1d5393cc6","Type":"ContainerStarted","Data":"cdf677629d8ef7dbffb372e1e320c66ffe36128c1655c99b0ff7036f3db68e4e"} Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.460152 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"28affc1b-492c-4671-b111-d2d1d5393cc6","Type":"ContainerStarted","Data":"821f41aaa1084afb405a53f0053ee33224073dcefdd86be2c0622d860919dedc"} Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.460168 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"28affc1b-492c-4671-b111-d2d1d5393cc6","Type":"ContainerStarted","Data":"459b0f030dac43b6b6e9a03039e3ce3ac85d62e623226079874e06b794a4d1cd"} Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.460519 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.464174 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558036-9tltz" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.464154 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558036-9tltz" event={"ID":"367c08b3-ceb2-4787-841b-c8bb6d3ce971","Type":"ContainerDied","Data":"d1b1644f1991b363a4fafeb2de745904c0856564ef87bd30e3185a21a2dd5fb1"} Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.464618 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1b1644f1991b363a4fafeb2de745904c0856564ef87bd30e3185a21a2dd5fb1" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.473833 4843 generic.go:334] "Generic (PLEG): container finished" podID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerID="a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9" exitCode=0 Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.473987 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerDied","Data":"a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9"} Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.474020 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dqj9j" event={"ID":"45d44dc1-12d7-422e-acbf-1a2f5008237b","Type":"ContainerDied","Data":"55ed977c5f95a97636006d6db113bbc6fd24dc1a81b0a5624ada19d8f534e2fc"} Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.474044 4843 scope.go:117] "RemoveContainer" containerID="a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.474054 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dqj9j" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.486754 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/watcher-kuttl-api-2" podStartSLOduration=2.48673871 podStartE2EDuration="2.48673871s" podCreationTimestamp="2026-03-14 09:56:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 09:56:04.486159176 +0000 UTC m=+2731.798770304" watchObservedRunningTime="2026-03-14 09:56:04.48673871 +0000 UTC m=+2731.799349828" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.513542 4843 scope.go:117] "RemoveContainer" containerID="0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.531693 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dqj9j"] Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.538048 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dqj9j"] Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.563853 4843 scope.go:117] "RemoveContainer" containerID="f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.597464 4843 scope.go:117] "RemoveContainer" containerID="a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9" Mar 14 09:56:04 crc kubenswrapper[4843]: E0314 09:56:04.597904 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9\": container with ID starting with a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9 not found: ID does not exist" containerID="a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.597950 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9"} err="failed to get container status \"a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9\": rpc error: code = NotFound desc = could not find container \"a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9\": container with ID starting with a86e8b6d45be1f93e37b8aa0bdb36eec01b102b23f122e0f5bd1813baee168e9 not found: ID does not exist" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.597984 4843 scope.go:117] "RemoveContainer" containerID="0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3" Mar 14 09:56:04 crc kubenswrapper[4843]: E0314 09:56:04.600640 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3\": container with ID starting with 0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3 not found: ID does not exist" containerID="0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.600685 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3"} err="failed to get container status \"0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3\": rpc error: code = NotFound desc = could not find container \"0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3\": container with ID starting with 0c1a784ef71eb5f152ca652d5f63d289ef24f83d8a8b53ca7fe7fac59c5a0bd3 not found: ID does not exist" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.600716 4843 scope.go:117] "RemoveContainer" containerID="f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e" Mar 14 09:56:04 crc kubenswrapper[4843]: E0314 09:56:04.601249 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e\": container with ID starting with f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e not found: ID does not exist" containerID="f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.601294 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e"} err="failed to get container status \"f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e\": rpc error: code = NotFound desc = could not find container \"f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e\": container with ID starting with f6651913442ed59cbd3c3ee56fd875c49e1228fda6cecfb317bd6e6f41f2505e not found: ID does not exist" Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.913057 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558030-x5ltd"] Mar 14 09:56:04 crc kubenswrapper[4843]: I0314 09:56:04.921778 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558030-x5ltd"] Mar 14 09:56:05 crc kubenswrapper[4843]: I0314 09:56:05.352428 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" path="/var/lib/kubelet/pods/45d44dc1-12d7-422e-acbf-1a2f5008237b/volumes" Mar 14 09:56:05 crc kubenswrapper[4843]: I0314 09:56:05.353898 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f61b3db5-0fa4-4809-b8b3-8974f6363775" path="/var/lib/kubelet/pods/f61b3db5-0fa4-4809-b8b3-8974f6363775/volumes" Mar 14 09:56:06 crc kubenswrapper[4843]: I0314 09:56:06.696030 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:07 crc kubenswrapper[4843]: I0314 09:56:07.831161 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.521080 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4tx56"] Mar 14 09:56:09 crc kubenswrapper[4843]: E0314 09:56:09.521819 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="extract-content" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.521837 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="extract-content" Mar 14 09:56:09 crc kubenswrapper[4843]: E0314 09:56:09.521868 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="registry-server" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.521875 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="registry-server" Mar 14 09:56:09 crc kubenswrapper[4843]: E0314 09:56:09.521893 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="extract-utilities" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.521902 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="extract-utilities" Mar 14 09:56:09 crc kubenswrapper[4843]: E0314 09:56:09.521915 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="367c08b3-ceb2-4787-841b-c8bb6d3ce971" containerName="oc" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.521922 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="367c08b3-ceb2-4787-841b-c8bb6d3ce971" containerName="oc" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.522105 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d44dc1-12d7-422e-acbf-1a2f5008237b" containerName="registry-server" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.522125 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="367c08b3-ceb2-4787-841b-c8bb6d3ce971" containerName="oc" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.523624 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.525646 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4tx56"] Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.681509 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvnt7\" (UniqueName: \"kubernetes.io/projected/9ff2f77c-0070-4b85-8890-5a76064582ec-kube-api-access-kvnt7\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.682598 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-catalog-content\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.682688 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-utilities\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.783989 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-utilities\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.784101 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvnt7\" (UniqueName: \"kubernetes.io/projected/9ff2f77c-0070-4b85-8890-5a76064582ec-kube-api-access-kvnt7\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.784138 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-catalog-content\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.784637 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-catalog-content\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.784972 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-utilities\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.825642 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvnt7\" (UniqueName: \"kubernetes.io/projected/9ff2f77c-0070-4b85-8890-5a76064582ec-kube-api-access-kvnt7\") pod \"certified-operators-4tx56\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:09 crc kubenswrapper[4843]: I0314 09:56:09.862345 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:10 crc kubenswrapper[4843]: I0314 09:56:10.452731 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4tx56"] Mar 14 09:56:10 crc kubenswrapper[4843]: I0314 09:56:10.543577 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerStarted","Data":"c05d3539c75f998d812f920de08cdb0c9ec1cfe8f0736e6e51f4f49b223edf44"} Mar 14 09:56:11 crc kubenswrapper[4843]: I0314 09:56:11.551633 4843 generic.go:334] "Generic (PLEG): container finished" podID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerID="001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27" exitCode=0 Mar 14 09:56:11 crc kubenswrapper[4843]: I0314 09:56:11.551679 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerDied","Data":"001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27"} Mar 14 09:56:12 crc kubenswrapper[4843]: I0314 09:56:12.563874 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerStarted","Data":"f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6"} Mar 14 09:56:12 crc kubenswrapper[4843]: I0314 09:56:12.830247 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:12 crc kubenswrapper[4843]: I0314 09:56:12.835187 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:13 crc kubenswrapper[4843]: I0314 09:56:13.574027 4843 generic.go:334] "Generic (PLEG): container finished" podID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerID="f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6" exitCode=0 Mar 14 09:56:13 crc kubenswrapper[4843]: I0314 09:56:13.574107 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerDied","Data":"f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6"} Mar 14 09:56:13 crc kubenswrapper[4843]: I0314 09:56:13.585848 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.139074 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.149489 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.149760 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-kuttl-api-log" containerID="cri-o://949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a" gracePeriod=30 Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.149819 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-1" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-api" containerID="cri-o://df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb" gracePeriod=30 Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.596977 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerStarted","Data":"ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475"} Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.600487 4843 generic.go:334] "Generic (PLEG): container finished" podID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerID="949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a" exitCode=143 Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.600554 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"94a9e1eb-8774-41d4-98fa-ea1112f21740","Type":"ContainerDied","Data":"949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a"} Mar 14 09:56:14 crc kubenswrapper[4843]: I0314 09:56:14.619193 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4tx56" podStartSLOduration=3.158272568 podStartE2EDuration="5.619168479s" podCreationTimestamp="2026-03-14 09:56:09 +0000 UTC" firstStartedPulling="2026-03-14 09:56:11.554263577 +0000 UTC m=+2738.866874705" lastFinishedPulling="2026-03-14 09:56:14.015159468 +0000 UTC m=+2741.327770616" observedRunningTime="2026-03-14 09:56:14.615528219 +0000 UTC m=+2741.928139347" watchObservedRunningTime="2026-03-14 09:56:14.619168479 +0000 UTC m=+2741.931779607" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.045373 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.084189 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-custom-prometheus-ca\") pod \"94a9e1eb-8774-41d4-98fa-ea1112f21740\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.084369 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-combined-ca-bundle\") pod \"94a9e1eb-8774-41d4-98fa-ea1112f21740\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.084427 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94a9e1eb-8774-41d4-98fa-ea1112f21740-logs\") pod \"94a9e1eb-8774-41d4-98fa-ea1112f21740\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.084455 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-config-data\") pod \"94a9e1eb-8774-41d4-98fa-ea1112f21740\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.084482 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-cert-memcached-mtls\") pod \"94a9e1eb-8774-41d4-98fa-ea1112f21740\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.084906 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a9e1eb-8774-41d4-98fa-ea1112f21740-logs" (OuterVolumeSpecName: "logs") pod "94a9e1eb-8774-41d4-98fa-ea1112f21740" (UID: "94a9e1eb-8774-41d4-98fa-ea1112f21740"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.085009 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lb68\" (UniqueName: \"kubernetes.io/projected/94a9e1eb-8774-41d4-98fa-ea1112f21740-kube-api-access-5lb68\") pod \"94a9e1eb-8774-41d4-98fa-ea1112f21740\" (UID: \"94a9e1eb-8774-41d4-98fa-ea1112f21740\") " Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.085489 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94a9e1eb-8774-41d4-98fa-ea1112f21740-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.127117 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94a9e1eb-8774-41d4-98fa-ea1112f21740" (UID: "94a9e1eb-8774-41d4-98fa-ea1112f21740"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.140726 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a9e1eb-8774-41d4-98fa-ea1112f21740-kube-api-access-5lb68" (OuterVolumeSpecName: "kube-api-access-5lb68") pod "94a9e1eb-8774-41d4-98fa-ea1112f21740" (UID: "94a9e1eb-8774-41d4-98fa-ea1112f21740"). InnerVolumeSpecName "kube-api-access-5lb68". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.188459 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lb68\" (UniqueName: \"kubernetes.io/projected/94a9e1eb-8774-41d4-98fa-ea1112f21740-kube-api-access-5lb68\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.188507 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.209437 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-config-data" (OuterVolumeSpecName: "config-data") pod "94a9e1eb-8774-41d4-98fa-ea1112f21740" (UID: "94a9e1eb-8774-41d4-98fa-ea1112f21740"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.225608 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "94a9e1eb-8774-41d4-98fa-ea1112f21740" (UID: "94a9e1eb-8774-41d4-98fa-ea1112f21740"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.295298 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.295339 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.311162 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "94a9e1eb-8774-41d4-98fa-ea1112f21740" (UID: "94a9e1eb-8774-41d4-98fa-ea1112f21740"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.387598 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.387649 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.387711 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.388356 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28e8d32901c9974c782e1e9aadfeb56ef673f0cf23ac2a95a7d8e3c463d945a9"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.388432 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://28e8d32901c9974c782e1e9aadfeb56ef673f0cf23ac2a95a7d8e3c463d945a9" gracePeriod=600 Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.397468 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/94a9e1eb-8774-41d4-98fa-ea1112f21740-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.613507 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="28e8d32901c9974c782e1e9aadfeb56ef673f0cf23ac2a95a7d8e3c463d945a9" exitCode=0 Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.613597 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"28e8d32901c9974c782e1e9aadfeb56ef673f0cf23ac2a95a7d8e3c463d945a9"} Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.613633 4843 scope.go:117] "RemoveContainer" containerID="5de64609b0e4fa11b27d408344ad6f0c003bb5ba620f27b32c630d25d9458aba" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.618857 4843 generic.go:334] "Generic (PLEG): container finished" podID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerID="df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb" exitCode=0 Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.619089 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-2" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-kuttl-api-log" containerID="cri-o://821f41aaa1084afb405a53f0053ee33224073dcefdd86be2c0622d860919dedc" gracePeriod=30 Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.619189 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-1" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.619424 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"94a9e1eb-8774-41d4-98fa-ea1112f21740","Type":"ContainerDied","Data":"df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb"} Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.619467 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-1" event={"ID":"94a9e1eb-8774-41d4-98fa-ea1112f21740","Type":"ContainerDied","Data":"2728517a9eb41ffdbc3a3dc492c1552f5cdbb345372d909a560331112f3da9bd"} Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.620604 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-2" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-api" containerID="cri-o://cdf677629d8ef7dbffb372e1e320c66ffe36128c1655c99b0ff7036f3db68e4e" gracePeriod=30 Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.654626 4843 scope.go:117] "RemoveContainer" containerID="df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.655740 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.666579 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-1"] Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.690589 4843 scope.go:117] "RemoveContainer" containerID="949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.718099 4843 scope.go:117] "RemoveContainer" containerID="df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb" Mar 14 09:56:15 crc kubenswrapper[4843]: E0314 09:56:15.719072 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb\": container with ID starting with df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb not found: ID does not exist" containerID="df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.719116 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb"} err="failed to get container status \"df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb\": rpc error: code = NotFound desc = could not find container \"df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb\": container with ID starting with df4bdefd995ce74cc167f08c1a0a4b01809927fb5219f6250fa7132186cd6efb not found: ID does not exist" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.719143 4843 scope.go:117] "RemoveContainer" containerID="949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a" Mar 14 09:56:15 crc kubenswrapper[4843]: E0314 09:56:15.719458 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a\": container with ID starting with 949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a not found: ID does not exist" containerID="949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a" Mar 14 09:56:15 crc kubenswrapper[4843]: I0314 09:56:15.719494 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a"} err="failed to get container status \"949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a\": rpc error: code = NotFound desc = could not find container \"949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a\": container with ID starting with 949bf8622965c8251f19e9cda6ad4e5cba5f386351140a63d748a9fc02de3e7a not found: ID does not exist" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.630068 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6"} Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.632627 4843 generic.go:334] "Generic (PLEG): container finished" podID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerID="cdf677629d8ef7dbffb372e1e320c66ffe36128c1655c99b0ff7036f3db68e4e" exitCode=0 Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.632663 4843 generic.go:334] "Generic (PLEG): container finished" podID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerID="821f41aaa1084afb405a53f0053ee33224073dcefdd86be2c0622d860919dedc" exitCode=143 Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.632693 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"28affc1b-492c-4671-b111-d2d1d5393cc6","Type":"ContainerDied","Data":"cdf677629d8ef7dbffb372e1e320c66ffe36128c1655c99b0ff7036f3db68e4e"} Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.632729 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"28affc1b-492c-4671-b111-d2d1d5393cc6","Type":"ContainerDied","Data":"821f41aaa1084afb405a53f0053ee33224073dcefdd86be2c0622d860919dedc"} Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.765192 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.921596 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-combined-ca-bundle\") pod \"28affc1b-492c-4671-b111-d2d1d5393cc6\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.921681 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28affc1b-492c-4671-b111-d2d1d5393cc6-logs\") pod \"28affc1b-492c-4671-b111-d2d1d5393cc6\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.921760 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m9gk\" (UniqueName: \"kubernetes.io/projected/28affc1b-492c-4671-b111-d2d1d5393cc6-kube-api-access-7m9gk\") pod \"28affc1b-492c-4671-b111-d2d1d5393cc6\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.921796 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-cert-memcached-mtls\") pod \"28affc1b-492c-4671-b111-d2d1d5393cc6\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.921866 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-custom-prometheus-ca\") pod \"28affc1b-492c-4671-b111-d2d1d5393cc6\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.921942 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-config-data\") pod \"28affc1b-492c-4671-b111-d2d1d5393cc6\" (UID: \"28affc1b-492c-4671-b111-d2d1d5393cc6\") " Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.922170 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28affc1b-492c-4671-b111-d2d1d5393cc6-logs" (OuterVolumeSpecName: "logs") pod "28affc1b-492c-4671-b111-d2d1d5393cc6" (UID: "28affc1b-492c-4671-b111-d2d1d5393cc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.923247 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28affc1b-492c-4671-b111-d2d1d5393cc6-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.926756 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28affc1b-492c-4671-b111-d2d1d5393cc6-kube-api-access-7m9gk" (OuterVolumeSpecName: "kube-api-access-7m9gk") pod "28affc1b-492c-4671-b111-d2d1d5393cc6" (UID: "28affc1b-492c-4671-b111-d2d1d5393cc6"). InnerVolumeSpecName "kube-api-access-7m9gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.954702 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28affc1b-492c-4671-b111-d2d1d5393cc6" (UID: "28affc1b-492c-4671-b111-d2d1d5393cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.975845 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-config-data" (OuterVolumeSpecName: "config-data") pod "28affc1b-492c-4671-b111-d2d1d5393cc6" (UID: "28affc1b-492c-4671-b111-d2d1d5393cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:16 crc kubenswrapper[4843]: I0314 09:56:16.984017 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "28affc1b-492c-4671-b111-d2d1d5393cc6" (UID: "28affc1b-492c-4671-b111-d2d1d5393cc6"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.008946 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "28affc1b-492c-4671-b111-d2d1d5393cc6" (UID: "28affc1b-492c-4671-b111-d2d1d5393cc6"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.024827 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.024855 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.024865 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.024874 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m9gk\" (UniqueName: \"kubernetes.io/projected/28affc1b-492c-4671-b111-d2d1d5393cc6-kube-api-access-7m9gk\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.024883 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/28affc1b-492c-4671-b111-d2d1d5393cc6-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.348032 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" path="/var/lib/kubelet/pods/94a9e1eb-8774-41d4-98fa-ea1112f21740/volumes" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.642156 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-2" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.642350 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-2" event={"ID":"28affc1b-492c-4671-b111-d2d1d5393cc6","Type":"ContainerDied","Data":"459b0f030dac43b6b6e9a03039e3ce3ac85d62e623226079874e06b794a4d1cd"} Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.642388 4843 scope.go:117] "RemoveContainer" containerID="cdf677629d8ef7dbffb372e1e320c66ffe36128c1655c99b0ff7036f3db68e4e" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.675863 4843 scope.go:117] "RemoveContainer" containerID="821f41aaa1084afb405a53f0053ee33224073dcefdd86be2c0622d860919dedc" Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.675876 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Mar 14 09:56:17 crc kubenswrapper[4843]: I0314 09:56:17.685045 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-2"] Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.503316 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.504072 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-kuttl-api-log" containerID="cri-o://01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632" gracePeriod=30 Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.504138 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-api" containerID="cri-o://7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27" gracePeriod=30 Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.652089 4843 generic.go:334] "Generic (PLEG): container finished" podID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerID="01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632" exitCode=143 Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.652170 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"843aaf0d-9d6e-486d-91b2-020cadb3b194","Type":"ContainerDied","Data":"01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632"} Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.986303 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-kuttl-api-log" probeResult="failure" output="Get \"http://10.217.1.13:9322/\": read tcp 10.217.0.2:36806->10.217.1.13:9322: read: connection reset by peer" Mar 14 09:56:18 crc kubenswrapper[4843]: I0314 09:56:18.986473 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/watcher-kuttl-api-0" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.1.13:9322/\": read tcp 10.217.0.2:36804->10.217.1.13:9322: read: connection reset by peer" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.359202 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" path="/var/lib/kubelet/pods/28affc1b-492c-4671-b111-d2d1d5393cc6/volumes" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.463248 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479105 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-config-data\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479145 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/843aaf0d-9d6e-486d-91b2-020cadb3b194-logs\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479294 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-cert-memcached-mtls\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479366 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479386 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-custom-prometheus-ca\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479463 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p59b\" (UniqueName: \"kubernetes.io/projected/843aaf0d-9d6e-486d-91b2-020cadb3b194-kube-api-access-7p59b\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.479795 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/843aaf0d-9d6e-486d-91b2-020cadb3b194-logs" (OuterVolumeSpecName: "logs") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.503072 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/843aaf0d-9d6e-486d-91b2-020cadb3b194-kube-api-access-7p59b" (OuterVolumeSpecName: "kube-api-access-7p59b") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "kube-api-access-7p59b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.537731 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.581013 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.582055 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle\") pod \"843aaf0d-9d6e-486d-91b2-020cadb3b194\" (UID: \"843aaf0d-9d6e-486d-91b2-020cadb3b194\") " Mar 14 09:56:19 crc kubenswrapper[4843]: W0314 09:56:19.582202 4843 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/843aaf0d-9d6e-486d-91b2-020cadb3b194/volumes/kubernetes.io~secret/combined-ca-bundle Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.582231 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.582434 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.582448 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.582458 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p59b\" (UniqueName: \"kubernetes.io/projected/843aaf0d-9d6e-486d-91b2-020cadb3b194-kube-api-access-7p59b\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.582467 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/843aaf0d-9d6e-486d-91b2-020cadb3b194-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.595732 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-config-data" (OuterVolumeSpecName: "config-data") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.604201 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "843aaf0d-9d6e-486d-91b2-020cadb3b194" (UID: "843aaf0d-9d6e-486d-91b2-020cadb3b194"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.669526 4843 generic.go:334] "Generic (PLEG): container finished" podID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerID="7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27" exitCode=0 Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.669582 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"843aaf0d-9d6e-486d-91b2-020cadb3b194","Type":"ContainerDied","Data":"7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27"} Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.669610 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-api-0" event={"ID":"843aaf0d-9d6e-486d-91b2-020cadb3b194","Type":"ContainerDied","Data":"c6b74c8bf8741ef504b611dac95dcb5d267359170d0e1ac614fab76048b55dc6"} Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.669645 4843 scope.go:117] "RemoveContainer" containerID="7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.669799 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-api-0" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.685420 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.685463 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/843aaf0d-9d6e-486d-91b2-020cadb3b194-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.715499 4843 scope.go:117] "RemoveContainer" containerID="01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.746345 4843 scope.go:117] "RemoveContainer" containerID="7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.748695 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.749945 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27\": container with ID starting with 7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27 not found: ID does not exist" containerID="7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.750002 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27"} err="failed to get container status \"7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27\": rpc error: code = NotFound desc = could not find container \"7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27\": container with ID starting with 7027c97275e069df79f6603f7d1c00ec35f77d28d86435b37c620edeeb34ca27 not found: ID does not exist" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.750036 4843 scope.go:117] "RemoveContainer" containerID="01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632" Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.753438 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632\": container with ID starting with 01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632 not found: ID does not exist" containerID="01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.753492 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632"} err="failed to get container status \"01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632\": rpc error: code = NotFound desc = could not find container \"01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632\": container with ID starting with 01a4ad63bebaa1e60bd1ff6eeabc5d5bb41ff0424733f608e111e72013426632 not found: ID does not exist" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.757778 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-api-0"] Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.772568 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n"] Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.784451 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-db-sync-bkc7n"] Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.824622 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.824865 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-applier-0" podUID="b2ce34bd-848a-451d-b698-cb9e208aec88" containerName="watcher-applier" containerID="cri-o://a121047c5cdf4ad35daa78b13fb464c5fa84f92d09ccd66907cdcde562dfa4a7" gracePeriod=30 Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.833502 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/watcherbacc-account-delete-s4fr8"] Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.833960 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.833985 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.834003 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834011 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.834029 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834040 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.834064 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834072 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.834088 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834095 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: E0314 09:56:19.834108 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834115 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834305 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834323 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834335 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a9e1eb-8774-41d4-98fa-ea1112f21740" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834345 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834355 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="28affc1b-492c-4671-b111-d2d1d5393cc6" containerName="watcher-api" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834364 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" containerName="watcher-kuttl-api-log" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.834941 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.861994 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherbacc-account-delete-s4fr8"] Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.863103 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.863663 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.868393 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.868581 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" podUID="0fb16236-e0bb-4b07-ae82-f10c19784b25" containerName="watcher-decision-engine" containerID="cri-o://09034172c7e541ed0e50b75dd991a61b4728706d7abc69c0f4cfb74c97d8b87a" gracePeriod=30 Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.888804 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1161e75-5892-4c93-9a72-d9709b619cca-operator-scripts\") pod \"watcherbacc-account-delete-s4fr8\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.888877 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxqwf\" (UniqueName: \"kubernetes.io/projected/c1161e75-5892-4c93-9a72-d9709b619cca-kube-api-access-sxqwf\") pod \"watcherbacc-account-delete-s4fr8\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.937785 4843 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.989520 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxqwf\" (UniqueName: \"kubernetes.io/projected/c1161e75-5892-4c93-9a72-d9709b619cca-kube-api-access-sxqwf\") pod \"watcherbacc-account-delete-s4fr8\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.989675 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1161e75-5892-4c93-9a72-d9709b619cca-operator-scripts\") pod \"watcherbacc-account-delete-s4fr8\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:19 crc kubenswrapper[4843]: I0314 09:56:19.990871 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1161e75-5892-4c93-9a72-d9709b619cca-operator-scripts\") pod \"watcherbacc-account-delete-s4fr8\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:20 crc kubenswrapper[4843]: I0314 09:56:20.028232 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxqwf\" (UniqueName: \"kubernetes.io/projected/c1161e75-5892-4c93-9a72-d9709b619cca-kube-api-access-sxqwf\") pod \"watcherbacc-account-delete-s4fr8\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:20 crc kubenswrapper[4843]: I0314 09:56:20.152334 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:20 crc kubenswrapper[4843]: I0314 09:56:20.684407 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/watcherbacc-account-delete-s4fr8"] Mar 14 09:56:20 crc kubenswrapper[4843]: I0314 09:56:20.733062 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:21 crc kubenswrapper[4843]: I0314 09:56:21.350514 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4083133e-f6cd-4ec6-8a41-61c440fa81c2" path="/var/lib/kubelet/pods/4083133e-f6cd-4ec6-8a41-61c440fa81c2/volumes" Mar 14 09:56:21 crc kubenswrapper[4843]: I0314 09:56:21.351527 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="843aaf0d-9d6e-486d-91b2-020cadb3b194" path="/var/lib/kubelet/pods/843aaf0d-9d6e-486d-91b2-020cadb3b194/volumes" Mar 14 09:56:21 crc kubenswrapper[4843]: I0314 09:56:21.690687 4843 generic.go:334] "Generic (PLEG): container finished" podID="c1161e75-5892-4c93-9a72-d9709b619cca" containerID="faaee28c2ddf305c67e11fee6328d5944040b414ad01ef9559666019cce0fa73" exitCode=0 Mar 14 09:56:21 crc kubenswrapper[4843]: I0314 09:56:21.690754 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" event={"ID":"c1161e75-5892-4c93-9a72-d9709b619cca","Type":"ContainerDied","Data":"faaee28c2ddf305c67e11fee6328d5944040b414ad01ef9559666019cce0fa73"} Mar 14 09:56:21 crc kubenswrapper[4843]: I0314 09:56:21.690792 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" event={"ID":"c1161e75-5892-4c93-9a72-d9709b619cca","Type":"ContainerStarted","Data":"b8920069f3ad61a5df27ff3aae6c042dacdd3499b39a27da0a4d6a5d6ae4147f"} Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.345729 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.346210 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-central-agent" containerID="cri-o://5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472" gracePeriod=30 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.347205 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="proxy-httpd" containerID="cri-o://4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50" gracePeriod=30 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.347411 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="sg-core" containerID="cri-o://19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c" gracePeriod=30 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.347482 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="watcher-kuttl-default/ceilometer-0" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-notification-agent" containerID="cri-o://eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9" gracePeriod=30 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.359709 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.16:3000/\": read tcp 10.217.0.2:56284->10.217.1.16:3000: read: connection reset by peer" Mar 14 09:56:22 crc kubenswrapper[4843]: E0314 09:56:22.433100 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2ce34bd_848a_451d_b698_cb9e208aec88.slice/crio-conmon-a121047c5cdf4ad35daa78b13fb464c5fa84f92d09ccd66907cdcde562dfa4a7.scope\": RecentStats: unable to find data in memory cache]" Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.701206 4843 generic.go:334] "Generic (PLEG): container finished" podID="b2ce34bd-848a-451d-b698-cb9e208aec88" containerID="a121047c5cdf4ad35daa78b13fb464c5fa84f92d09ccd66907cdcde562dfa4a7" exitCode=0 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.701302 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2ce34bd-848a-451d-b698-cb9e208aec88","Type":"ContainerDied","Data":"a121047c5cdf4ad35daa78b13fb464c5fa84f92d09ccd66907cdcde562dfa4a7"} Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.704999 4843 generic.go:334] "Generic (PLEG): container finished" podID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerID="4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50" exitCode=0 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.705032 4843 generic.go:334] "Generic (PLEG): container finished" podID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerID="19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c" exitCode=2 Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.705123 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerDied","Data":"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50"} Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.706982 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerDied","Data":"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c"} Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.773263 4843 prober.go:107] "Probe failed" probeType="Readiness" pod="watcher-kuttl-default/ceilometer-0" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.16:3000/\": dial tcp 10.217.1.16:3000: connect: connection refused" Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.891781 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:56:22 crc kubenswrapper[4843]: I0314 09:56:22.988330 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.041707 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2ce34bd-848a-451d-b698-cb9e208aec88-logs\") pod \"b2ce34bd-848a-451d-b698-cb9e208aec88\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.042595 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p55m4\" (UniqueName: \"kubernetes.io/projected/b2ce34bd-848a-451d-b698-cb9e208aec88-kube-api-access-p55m4\") pod \"b2ce34bd-848a-451d-b698-cb9e208aec88\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.042502 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ce34bd-848a-451d-b698-cb9e208aec88-logs" (OuterVolumeSpecName: "logs") pod "b2ce34bd-848a-451d-b698-cb9e208aec88" (UID: "b2ce34bd-848a-451d-b698-cb9e208aec88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.042982 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-combined-ca-bundle\") pod \"b2ce34bd-848a-451d-b698-cb9e208aec88\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.043089 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-cert-memcached-mtls\") pod \"b2ce34bd-848a-451d-b698-cb9e208aec88\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.043190 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxqwf\" (UniqueName: \"kubernetes.io/projected/c1161e75-5892-4c93-9a72-d9709b619cca-kube-api-access-sxqwf\") pod \"c1161e75-5892-4c93-9a72-d9709b619cca\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.043316 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1161e75-5892-4c93-9a72-d9709b619cca-operator-scripts\") pod \"c1161e75-5892-4c93-9a72-d9709b619cca\" (UID: \"c1161e75-5892-4c93-9a72-d9709b619cca\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.043441 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-config-data\") pod \"b2ce34bd-848a-451d-b698-cb9e208aec88\" (UID: \"b2ce34bd-848a-451d-b698-cb9e208aec88\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.043991 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2ce34bd-848a-451d-b698-cb9e208aec88-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.046381 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1161e75-5892-4c93-9a72-d9709b619cca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1161e75-5892-4c93-9a72-d9709b619cca" (UID: "c1161e75-5892-4c93-9a72-d9709b619cca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.048399 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ce34bd-848a-451d-b698-cb9e208aec88-kube-api-access-p55m4" (OuterVolumeSpecName: "kube-api-access-p55m4") pod "b2ce34bd-848a-451d-b698-cb9e208aec88" (UID: "b2ce34bd-848a-451d-b698-cb9e208aec88"). InnerVolumeSpecName "kube-api-access-p55m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.049489 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1161e75-5892-4c93-9a72-d9709b619cca-kube-api-access-sxqwf" (OuterVolumeSpecName: "kube-api-access-sxqwf") pod "c1161e75-5892-4c93-9a72-d9709b619cca" (UID: "c1161e75-5892-4c93-9a72-d9709b619cca"). InnerVolumeSpecName "kube-api-access-sxqwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.071742 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2ce34bd-848a-451d-b698-cb9e208aec88" (UID: "b2ce34bd-848a-451d-b698-cb9e208aec88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.093385 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-config-data" (OuterVolumeSpecName: "config-data") pod "b2ce34bd-848a-451d-b698-cb9e208aec88" (UID: "b2ce34bd-848a-451d-b698-cb9e208aec88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.101455 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.126365 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "b2ce34bd-848a-451d-b698-cb9e208aec88" (UID: "b2ce34bd-848a-451d-b698-cb9e208aec88"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145448 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-scripts\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145550 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-run-httpd\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145652 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-log-httpd\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145694 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-ceilometer-tls-certs\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145727 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-sg-core-conf-yaml\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145755 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-combined-ca-bundle\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145779 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-config-data\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.145802 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctrkm\" (UniqueName: \"kubernetes.io/projected/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-kube-api-access-ctrkm\") pod \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\" (UID: \"00e2f129-cb1f-48d7-bb47-603ec1bbdec7\") " Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146037 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146082 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146097 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146112 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxqwf\" (UniqueName: \"kubernetes.io/projected/c1161e75-5892-4c93-9a72-d9709b619cca-kube-api-access-sxqwf\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146125 4843 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1161e75-5892-4c93-9a72-d9709b619cca-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146140 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2ce34bd-848a-451d-b698-cb9e208aec88-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146150 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p55m4\" (UniqueName: \"kubernetes.io/projected/b2ce34bd-848a-451d-b698-cb9e208aec88-kube-api-access-p55m4\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.146856 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.158454 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-scripts" (OuterVolumeSpecName: "scripts") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.163539 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-kube-api-access-ctrkm" (OuterVolumeSpecName: "kube-api-access-ctrkm") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "kube-api-access-ctrkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.196026 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.206027 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.248509 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctrkm\" (UniqueName: \"kubernetes.io/projected/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-kube-api-access-ctrkm\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.248546 4843 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-scripts\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.248560 4843 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.248572 4843 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.248583 4843 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.248594 4843 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.249531 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.257715 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-config-data" (OuterVolumeSpecName: "config-data") pod "00e2f129-cb1f-48d7-bb47-603ec1bbdec7" (UID: "00e2f129-cb1f-48d7-bb47-603ec1bbdec7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.349662 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.358442 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e2f129-cb1f-48d7-bb47-603ec1bbdec7-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.500020 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4tx56"] Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.713914 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-applier-0" event={"ID":"b2ce34bd-848a-451d-b698-cb9e208aec88","Type":"ContainerDied","Data":"db495bd932d770248f5787cf5eeb8cf85e3d612bee49a5a61c5d795e659b168a"} Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.714178 4843 scope.go:117] "RemoveContainer" containerID="a121047c5cdf4ad35daa78b13fb464c5fa84f92d09ccd66907cdcde562dfa4a7" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.713968 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-applier-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.716679 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.716747 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcherbacc-account-delete-s4fr8" event={"ID":"c1161e75-5892-4c93-9a72-d9709b619cca","Type":"ContainerDied","Data":"b8920069f3ad61a5df27ff3aae6c042dacdd3499b39a27da0a4d6a5d6ae4147f"} Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.716787 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8920069f3ad61a5df27ff3aae6c042dacdd3499b39a27da0a4d6a5d6ae4147f" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.720523 4843 generic.go:334] "Generic (PLEG): container finished" podID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerID="eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9" exitCode=0 Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.720546 4843 generic.go:334] "Generic (PLEG): container finished" podID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerID="5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472" exitCode=0 Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.720568 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerDied","Data":"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9"} Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.720588 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.720609 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerDied","Data":"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472"} Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.720623 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"00e2f129-cb1f-48d7-bb47-603ec1bbdec7","Type":"ContainerDied","Data":"33cfed407eae5920a287b73b7758495fd772838906f797dc7b04491e2994bb3a"} Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.722414 4843 generic.go:334] "Generic (PLEG): container finished" podID="0fb16236-e0bb-4b07-ae82-f10c19784b25" containerID="09034172c7e541ed0e50b75dd991a61b4728706d7abc69c0f4cfb74c97d8b87a" exitCode=0 Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.722497 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0fb16236-e0bb-4b07-ae82-f10c19784b25","Type":"ContainerDied","Data":"09034172c7e541ed0e50b75dd991a61b4728706d7abc69c0f4cfb74c97d8b87a"} Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.722622 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4tx56" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="registry-server" containerID="cri-o://ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475" gracePeriod=2 Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.737795 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.744110 4843 scope.go:117] "RemoveContainer" containerID="4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.744294 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-applier-0"] Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.761006 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.782000 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.793816 4843 scope.go:117] "RemoveContainer" containerID="19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.835513 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:56:23 crc kubenswrapper[4843]: E0314 09:56:23.835948 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-notification-agent" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.835972 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-notification-agent" Mar 14 09:56:23 crc kubenswrapper[4843]: E0314 09:56:23.835991 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="proxy-httpd" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836001 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="proxy-httpd" Mar 14 09:56:23 crc kubenswrapper[4843]: E0314 09:56:23.836018 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ce34bd-848a-451d-b698-cb9e208aec88" containerName="watcher-applier" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836028 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ce34bd-848a-451d-b698-cb9e208aec88" containerName="watcher-applier" Mar 14 09:56:23 crc kubenswrapper[4843]: E0314 09:56:23.836046 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1161e75-5892-4c93-9a72-d9709b619cca" containerName="mariadb-account-delete" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836054 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1161e75-5892-4c93-9a72-d9709b619cca" containerName="mariadb-account-delete" Mar 14 09:56:23 crc kubenswrapper[4843]: E0314 09:56:23.836065 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-central-agent" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836073 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-central-agent" Mar 14 09:56:23 crc kubenswrapper[4843]: E0314 09:56:23.836095 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="sg-core" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836104 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="sg-core" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836325 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-notification-agent" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836341 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="proxy-httpd" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836367 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="ceilometer-central-agent" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836379 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" containerName="sg-core" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836391 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ce34bd-848a-451d-b698-cb9e208aec88" containerName="watcher-applier" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.836405 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1161e75-5892-4c93-9a72-d9709b619cca" containerName="mariadb-account-delete" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.838212 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.843837 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.847774 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-config-data" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.848026 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"cert-ceilometer-internal-svc" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.848840 4843 reflector.go:368] Caches populated for *v1.Secret from object-"watcher-kuttl-default"/"ceilometer-scripts" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867294 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b50529a-c274-4664-b0ef-a06045bbfb96-log-httpd\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867359 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-config-data\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867433 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz9k4\" (UniqueName: \"kubernetes.io/projected/0b50529a-c274-4664-b0ef-a06045bbfb96-kube-api-access-nz9k4\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867499 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867548 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-scripts\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867574 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b50529a-c274-4664-b0ef-a06045bbfb96-run-httpd\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867594 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.867618 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.927138 4843 scope.go:117] "RemoveContainer" containerID="eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.954233 4843 scope.go:117] "RemoveContainer" containerID="5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972103 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-scripts\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972152 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b50529a-c274-4664-b0ef-a06045bbfb96-run-httpd\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972181 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972211 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972247 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b50529a-c274-4664-b0ef-a06045bbfb96-log-httpd\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972307 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-config-data\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972361 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz9k4\" (UniqueName: \"kubernetes.io/projected/0b50529a-c274-4664-b0ef-a06045bbfb96-kube-api-access-nz9k4\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972401 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972737 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b50529a-c274-4664-b0ef-a06045bbfb96-run-httpd\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.972985 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b50529a-c274-4664-b0ef-a06045bbfb96-log-httpd\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.981071 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.981415 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.984849 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-scripts\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:23 crc kubenswrapper[4843]: I0314 09:56:23.984865 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-config-data\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:23.999521 4843 scope.go:117] "RemoveContainer" containerID="4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.000737 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50\": container with ID starting with 4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50 not found: ID does not exist" containerID="4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.000802 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50"} err="failed to get container status \"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50\": rpc error: code = NotFound desc = could not find container \"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50\": container with ID starting with 4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.000834 4843 scope.go:117] "RemoveContainer" containerID="19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.001698 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c\": container with ID starting with 19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c not found: ID does not exist" containerID="19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.001745 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c"} err="failed to get container status \"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c\": rpc error: code = NotFound desc = could not find container \"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c\": container with ID starting with 19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.001764 4843 scope.go:117] "RemoveContainer" containerID="eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.002020 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9\": container with ID starting with eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9 not found: ID does not exist" containerID="eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002045 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9"} err="failed to get container status \"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9\": rpc error: code = NotFound desc = could not find container \"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9\": container with ID starting with eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002062 4843 scope.go:117] "RemoveContainer" containerID="5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.002288 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472\": container with ID starting with 5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472 not found: ID does not exist" containerID="5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002311 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472"} err="failed to get container status \"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472\": rpc error: code = NotFound desc = could not find container \"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472\": container with ID starting with 5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002329 4843 scope.go:117] "RemoveContainer" containerID="4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002536 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50"} err="failed to get container status \"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50\": rpc error: code = NotFound desc = could not find container \"4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50\": container with ID starting with 4e427f0af326985ff2e8f0f36f46faf7651e99cf9b1c3de801900ef3ed8a5d50 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002555 4843 scope.go:117] "RemoveContainer" containerID="19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002754 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c"} err="failed to get container status \"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c\": rpc error: code = NotFound desc = could not find container \"19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c\": container with ID starting with 19a41b8de3f00645ea0b6188609a7fd2166e245866b0400673946ccc1bb8732c not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.002779 4843 scope.go:117] "RemoveContainer" containerID="eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.003125 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9"} err="failed to get container status \"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9\": rpc error: code = NotFound desc = could not find container \"eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9\": container with ID starting with eb16f720fee5779b51b0716a4cccb572922e6838590e37e343472b62c4552fa9 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.003162 4843 scope.go:117] "RemoveContainer" containerID="5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.003385 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472"} err="failed to get container status \"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472\": rpc error: code = NotFound desc = could not find container \"5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472\": container with ID starting with 5214a19e100411779ce34dae1b1d706c314061d5a536def4773499bf285a8472 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.004362 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b50529a-c274-4664-b0ef-a06045bbfb96-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.005389 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz9k4\" (UniqueName: \"kubernetes.io/projected/0b50529a-c274-4664-b0ef-a06045bbfb96-kube-api-access-nz9k4\") pod \"ceilometer-0\" (UID: \"0b50529a-c274-4664-b0ef-a06045bbfb96\") " pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.211293 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.234015 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.241875 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.291998 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb16236-e0bb-4b07-ae82-f10c19784b25-logs\") pod \"0fb16236-e0bb-4b07-ae82-f10c19784b25\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292065 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-combined-ca-bundle\") pod \"0fb16236-e0bb-4b07-ae82-f10c19784b25\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292096 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq9pk\" (UniqueName: \"kubernetes.io/projected/0fb16236-e0bb-4b07-ae82-f10c19784b25-kube-api-access-bq9pk\") pod \"0fb16236-e0bb-4b07-ae82-f10c19784b25\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292138 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-custom-prometheus-ca\") pod \"0fb16236-e0bb-4b07-ae82-f10c19784b25\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292260 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-utilities\") pod \"9ff2f77c-0070-4b85-8890-5a76064582ec\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292310 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-cert-memcached-mtls\") pod \"0fb16236-e0bb-4b07-ae82-f10c19784b25\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292339 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-config-data\") pod \"0fb16236-e0bb-4b07-ae82-f10c19784b25\" (UID: \"0fb16236-e0bb-4b07-ae82-f10c19784b25\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292364 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-catalog-content\") pod \"9ff2f77c-0070-4b85-8890-5a76064582ec\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.292409 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvnt7\" (UniqueName: \"kubernetes.io/projected/9ff2f77c-0070-4b85-8890-5a76064582ec-kube-api-access-kvnt7\") pod \"9ff2f77c-0070-4b85-8890-5a76064582ec\" (UID: \"9ff2f77c-0070-4b85-8890-5a76064582ec\") " Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.297935 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb16236-e0bb-4b07-ae82-f10c19784b25-logs" (OuterVolumeSpecName: "logs") pod "0fb16236-e0bb-4b07-ae82-f10c19784b25" (UID: "0fb16236-e0bb-4b07-ae82-f10c19784b25"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.301211 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-utilities" (OuterVolumeSpecName: "utilities") pod "9ff2f77c-0070-4b85-8890-5a76064582ec" (UID: "9ff2f77c-0070-4b85-8890-5a76064582ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.303412 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb16236-e0bb-4b07-ae82-f10c19784b25-kube-api-access-bq9pk" (OuterVolumeSpecName: "kube-api-access-bq9pk") pod "0fb16236-e0bb-4b07-ae82-f10c19784b25" (UID: "0fb16236-e0bb-4b07-ae82-f10c19784b25"). InnerVolumeSpecName "kube-api-access-bq9pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.307491 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff2f77c-0070-4b85-8890-5a76064582ec-kube-api-access-kvnt7" (OuterVolumeSpecName: "kube-api-access-kvnt7") pod "9ff2f77c-0070-4b85-8890-5a76064582ec" (UID: "9ff2f77c-0070-4b85-8890-5a76064582ec"). InnerVolumeSpecName "kube-api-access-kvnt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.336421 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fb16236-e0bb-4b07-ae82-f10c19784b25" (UID: "0fb16236-e0bb-4b07-ae82-f10c19784b25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.340333 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "0fb16236-e0bb-4b07-ae82-f10c19784b25" (UID: "0fb16236-e0bb-4b07-ae82-f10c19784b25"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.368550 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-config-data" (OuterVolumeSpecName: "config-data") pod "0fb16236-e0bb-4b07-ae82-f10c19784b25" (UID: "0fb16236-e0bb-4b07-ae82-f10c19784b25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.384333 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ff2f77c-0070-4b85-8890-5a76064582ec" (UID: "9ff2f77c-0070-4b85-8890-5a76064582ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394572 4843 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-utilities\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394603 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394619 4843 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ff2f77c-0070-4b85-8890-5a76064582ec-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394633 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvnt7\" (UniqueName: \"kubernetes.io/projected/9ff2f77c-0070-4b85-8890-5a76064582ec-kube-api-access-kvnt7\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394644 4843 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb16236-e0bb-4b07-ae82-f10c19784b25-logs\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394655 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394667 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq9pk\" (UniqueName: \"kubernetes.io/projected/0fb16236-e0bb-4b07-ae82-f10c19784b25-kube-api-access-bq9pk\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.394676 4843 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.413402 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "0fb16236-e0bb-4b07-ae82-f10c19784b25" (UID: "0fb16236-e0bb-4b07-ae82-f10c19784b25"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.495675 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/0fb16236-e0bb-4b07-ae82-f10c19784b25-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.682029 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/ceilometer-0"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.731289 4843 generic.go:334] "Generic (PLEG): container finished" podID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerID="ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475" exitCode=0 Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.731347 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerDied","Data":"ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475"} Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.731376 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4tx56" event={"ID":"9ff2f77c-0070-4b85-8890-5a76064582ec","Type":"ContainerDied","Data":"c05d3539c75f998d812f920de08cdb0c9ec1cfe8f0736e6e51f4f49b223edf44"} Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.731390 4843 scope.go:117] "RemoveContainer" containerID="ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.731512 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4tx56" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.753325 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" event={"ID":"0fb16236-e0bb-4b07-ae82-f10c19784b25","Type":"ContainerDied","Data":"60e1cd04eda923dbfacdcb061dea421b85de46f713a04ce833e9c5b362bb6ab0"} Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.753379 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/watcher-kuttl-decision-engine-0" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.754761 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0b50529a-c274-4664-b0ef-a06045bbfb96","Type":"ContainerStarted","Data":"da4e4e68efd1ae993a2b0bf34a8f9e54b9977a96fc3931c972bcaa8920879838"} Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.770148 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4tx56"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.770502 4843 scope.go:117] "RemoveContainer" containerID="f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.790473 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4tx56"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.793584 4843 scope.go:117] "RemoveContainer" containerID="001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.799201 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.810563 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-kuttl-decision-engine-0"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.841411 4843 scope.go:117] "RemoveContainer" containerID="ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.842159 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475\": container with ID starting with ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475 not found: ID does not exist" containerID="ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.842201 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475"} err="failed to get container status \"ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475\": rpc error: code = NotFound desc = could not find container \"ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475\": container with ID starting with ce7af43a522a1eb9a3d7e90c0e1e69d54cca2d587ad642146fa0959ca9c0d475 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.842228 4843 scope.go:117] "RemoveContainer" containerID="f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.842672 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6\": container with ID starting with f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6 not found: ID does not exist" containerID="f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.842702 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6"} err="failed to get container status \"f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6\": rpc error: code = NotFound desc = could not find container \"f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6\": container with ID starting with f7c817ca17f72b42507dec7a4d58c3d61f91cb3fe26b63cab2ec6d7b7fef00d6 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.842723 4843 scope.go:117] "RemoveContainer" containerID="001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27" Mar 14 09:56:24 crc kubenswrapper[4843]: E0314 09:56:24.846533 4843 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27\": container with ID starting with 001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27 not found: ID does not exist" containerID="001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.846585 4843 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27"} err="failed to get container status \"001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27\": rpc error: code = NotFound desc = could not find container \"001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27\": container with ID starting with 001d64f150e1a0a733617d1f97ec0d9ffd66991a0e87b4a548bac4b19dee9a27 not found: ID does not exist" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.846618 4843 scope.go:117] "RemoveContainer" containerID="09034172c7e541ed0e50b75dd991a61b4728706d7abc69c0f4cfb74c97d8b87a" Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.846762 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-db-create-xp85v"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.864879 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-db-create-xp85v"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.872120 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcher-bacc-account-create-update-qzv47"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.879711 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcher-bacc-account-create-update-qzv47"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.887266 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/watcherbacc-account-delete-s4fr8"] Mar 14 09:56:24 crc kubenswrapper[4843]: I0314 09:56:24.893469 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/watcherbacc-account-delete-s4fr8"] Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.352969 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00e2f129-cb1f-48d7-bb47-603ec1bbdec7" path="/var/lib/kubelet/pods/00e2f129-cb1f-48d7-bb47-603ec1bbdec7/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.353822 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5423a2-a865-4402-90b2-9053439f187b" path="/var/lib/kubelet/pods/0c5423a2-a865-4402-90b2-9053439f187b/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.354486 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fb16236-e0bb-4b07-ae82-f10c19784b25" path="/var/lib/kubelet/pods/0fb16236-e0bb-4b07-ae82-f10c19784b25/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.355463 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" path="/var/lib/kubelet/pods/9ff2f77c-0070-4b85-8890-5a76064582ec/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.356438 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ce34bd-848a-451d-b698-cb9e208aec88" path="/var/lib/kubelet/pods/b2ce34bd-848a-451d-b698-cb9e208aec88/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.357025 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f7a856-c96c-4553-84d8-53e419c1d015" path="/var/lib/kubelet/pods/b6f7a856-c96c-4553-84d8-53e419c1d015/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.358182 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1161e75-5892-4c93-9a72-d9709b619cca" path="/var/lib/kubelet/pods/c1161e75-5892-4c93-9a72-d9709b619cca/volumes" Mar 14 09:56:25 crc kubenswrapper[4843]: I0314 09:56:25.769417 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0b50529a-c274-4664-b0ef-a06045bbfb96","Type":"ContainerStarted","Data":"ea96f5d95a45ec77e8dbe4399c476e8eff5cbb3b52284070e871ecd3145c9861"} Mar 14 09:56:26 crc kubenswrapper[4843]: I0314 09:56:26.788462 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0b50529a-c274-4664-b0ef-a06045bbfb96","Type":"ContainerStarted","Data":"bcd7f1ea07ee11accaa56bae9b1a285fd854b0ac8eb23a4947647dd25e3a5ccb"} Mar 14 09:56:27 crc kubenswrapper[4843]: I0314 09:56:27.799077 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0b50529a-c274-4664-b0ef-a06045bbfb96","Type":"ContainerStarted","Data":"f8ad0eb79592d0866f86e11a3789be6aab230b23be6c501047c86b241a9476ff"} Mar 14 09:56:28 crc kubenswrapper[4843]: I0314 09:56:28.809910 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/ceilometer-0" event={"ID":"0b50529a-c274-4664-b0ef-a06045bbfb96","Type":"ContainerStarted","Data":"b3190e2d515cb4e7aa45294722eed38bff00dbf369a7165ae971a06cdbdcb34d"} Mar 14 09:56:28 crc kubenswrapper[4843]: I0314 09:56:28.810234 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:28 crc kubenswrapper[4843]: I0314 09:56:28.857443 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/ceilometer-0" podStartSLOduration=2.423465841 podStartE2EDuration="5.85742169s" podCreationTimestamp="2026-03-14 09:56:23 +0000 UTC" firstStartedPulling="2026-03-14 09:56:24.683633574 +0000 UTC m=+2751.996244702" lastFinishedPulling="2026-03-14 09:56:28.117589423 +0000 UTC m=+2755.430200551" observedRunningTime="2026-03-14 09:56:28.852642143 +0000 UTC m=+2756.165253301" watchObservedRunningTime="2026-03-14 09:56:28.85742169 +0000 UTC m=+2756.170032818" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.984265 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s74vk/must-gather-88lrc"] Mar 14 09:56:48 crc kubenswrapper[4843]: E0314 09:56:48.986130 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="extract-utilities" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.986224 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="extract-utilities" Mar 14 09:56:48 crc kubenswrapper[4843]: E0314 09:56:48.986252 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="extract-content" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.986261 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="extract-content" Mar 14 09:56:48 crc kubenswrapper[4843]: E0314 09:56:48.986297 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="registry-server" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.986306 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="registry-server" Mar 14 09:56:48 crc kubenswrapper[4843]: E0314 09:56:48.986318 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb16236-e0bb-4b07-ae82-f10c19784b25" containerName="watcher-decision-engine" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.986325 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb16236-e0bb-4b07-ae82-f10c19784b25" containerName="watcher-decision-engine" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.986509 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ff2f77c-0070-4b85-8890-5a76064582ec" containerName="registry-server" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.986522 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb16236-e0bb-4b07-ae82-f10c19784b25" containerName="watcher-decision-engine" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.987760 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.990290 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s74vk"/"kube-root-ca.crt" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.991160 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-s74vk"/"default-dockercfg-ss8td" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.994708 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s74vk"/"openshift-service-ca.crt" Mar 14 09:56:48 crc kubenswrapper[4843]: I0314 09:56:48.994897 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s74vk/must-gather-88lrc"] Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.100336 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4c313e64-65bc-4dfd-a1cf-61956ec5870e-must-gather-output\") pod \"must-gather-88lrc\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.100402 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmwx4\" (UniqueName: \"kubernetes.io/projected/4c313e64-65bc-4dfd-a1cf-61956ec5870e-kube-api-access-jmwx4\") pod \"must-gather-88lrc\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.202356 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4c313e64-65bc-4dfd-a1cf-61956ec5870e-must-gather-output\") pod \"must-gather-88lrc\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.202659 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmwx4\" (UniqueName: \"kubernetes.io/projected/4c313e64-65bc-4dfd-a1cf-61956ec5870e-kube-api-access-jmwx4\") pod \"must-gather-88lrc\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.202915 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4c313e64-65bc-4dfd-a1cf-61956ec5870e-must-gather-output\") pod \"must-gather-88lrc\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.225024 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmwx4\" (UniqueName: \"kubernetes.io/projected/4c313e64-65bc-4dfd-a1cf-61956ec5870e-kube-api-access-jmwx4\") pod \"must-gather-88lrc\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.315510 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.914815 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s74vk/must-gather-88lrc"] Mar 14 09:56:49 crc kubenswrapper[4843]: I0314 09:56:49.988401 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s74vk/must-gather-88lrc" event={"ID":"4c313e64-65bc-4dfd-a1cf-61956ec5870e","Type":"ContainerStarted","Data":"2a8d2d2329d1c70850c76194d1db99bd7661e8d3827bd13589eb65d454950248"} Mar 14 09:56:52 crc kubenswrapper[4843]: I0314 09:56:52.270662 4843 scope.go:117] "RemoveContainer" containerID="08441fe3385a19c95b327c82bd4507f5309a6a031a84ec3e29c38a1ecd019cd5" Mar 14 09:56:54 crc kubenswrapper[4843]: I0314 09:56:54.309911 4843 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="watcher-kuttl-default/ceilometer-0" Mar 14 09:56:56 crc kubenswrapper[4843]: I0314 09:56:56.275928 4843 scope.go:117] "RemoveContainer" containerID="1c6382b86c92b9d70e09cd01caa74c0fb3fad47108f970198fbe644e2922aae9" Mar 14 09:56:56 crc kubenswrapper[4843]: I0314 09:56:56.317244 4843 scope.go:117] "RemoveContainer" containerID="afdb4c8617528309845b42d2dbccbb9275fed12ae0586ad2c08b3ec68f467a31" Mar 14 09:56:56 crc kubenswrapper[4843]: I0314 09:56:56.338533 4843 scope.go:117] "RemoveContainer" containerID="0630c13543895bc2d1f4e511d3e745e8bbb62de1b1cee339f0e19496de0d8aee" Mar 14 09:56:56 crc kubenswrapper[4843]: I0314 09:56:56.354718 4843 scope.go:117] "RemoveContainer" containerID="d9f8c36fbcef7e13f7b5aea054b0505e7e72074e8413cfd08b8d8d392a084f49" Mar 14 09:56:57 crc kubenswrapper[4843]: I0314 09:56:57.087001 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s74vk/must-gather-88lrc" event={"ID":"4c313e64-65bc-4dfd-a1cf-61956ec5870e","Type":"ContainerStarted","Data":"2dc2d95ed3d84c4fe62cd7638b6419f3992efdffa94360126f8f6def20ba592c"} Mar 14 09:56:57 crc kubenswrapper[4843]: I0314 09:56:57.087214 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s74vk/must-gather-88lrc" event={"ID":"4c313e64-65bc-4dfd-a1cf-61956ec5870e","Type":"ContainerStarted","Data":"bac7c0be5cf494e8c6e4d713e9b7c582f3278fb20ce89aae26628efc5a628cf6"} Mar 14 09:56:57 crc kubenswrapper[4843]: I0314 09:56:57.114174 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s74vk/must-gather-88lrc" podStartSLOduration=2.695735868 podStartE2EDuration="9.114149249s" podCreationTimestamp="2026-03-14 09:56:48 +0000 UTC" firstStartedPulling="2026-03-14 09:56:49.924937933 +0000 UTC m=+2777.237549061" lastFinishedPulling="2026-03-14 09:56:56.343351304 +0000 UTC m=+2783.655962442" observedRunningTime="2026-03-14 09:56:57.111507384 +0000 UTC m=+2784.424118512" watchObservedRunningTime="2026-03-14 09:56:57.114149249 +0000 UTC m=+2784.426760387" Mar 14 09:57:56 crc kubenswrapper[4843]: I0314 09:57:56.551586 4843 scope.go:117] "RemoveContainer" containerID="3c6e0df8928364da3119289d1ba8b72295a281e8e99c8f0fefc0ea531db15869" Mar 14 09:57:56 crc kubenswrapper[4843]: I0314 09:57:56.574633 4843 scope.go:117] "RemoveContainer" containerID="50851d18785e26c731d169912ceb516dd101c013eebe32bd01d97b47ded00d0a" Mar 14 09:57:56 crc kubenswrapper[4843]: I0314 09:57:56.633454 4843 scope.go:117] "RemoveContainer" containerID="bc0ebf7333899aae0cbf31cfc8a0453b4462315bb665ab7383b5c06f1b08be90" Mar 14 09:57:56 crc kubenswrapper[4843]: I0314 09:57:56.664227 4843 scope.go:117] "RemoveContainer" containerID="4a2a26597a24034cd40558c7ee55ba24a97b59c22547df9bdf88068307b14065" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.141183 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558038-thp2w"] Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.142751 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.145412 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.145671 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.146054 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.159762 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558038-thp2w"] Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.306334 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p985w\" (UniqueName: \"kubernetes.io/projected/c6fea642-5dda-4baa-8b0e-9fe644a58a2f-kube-api-access-p985w\") pod \"auto-csr-approver-29558038-thp2w\" (UID: \"c6fea642-5dda-4baa-8b0e-9fe644a58a2f\") " pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.407954 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p985w\" (UniqueName: \"kubernetes.io/projected/c6fea642-5dda-4baa-8b0e-9fe644a58a2f-kube-api-access-p985w\") pod \"auto-csr-approver-29558038-thp2w\" (UID: \"c6fea642-5dda-4baa-8b0e-9fe644a58a2f\") " pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.430949 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p985w\" (UniqueName: \"kubernetes.io/projected/c6fea642-5dda-4baa-8b0e-9fe644a58a2f-kube-api-access-p985w\") pod \"auto-csr-approver-29558038-thp2w\" (UID: \"c6fea642-5dda-4baa-8b0e-9fe644a58a2f\") " pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.457841 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:00 crc kubenswrapper[4843]: I0314 09:58:00.973602 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558038-thp2w"] Mar 14 09:58:01 crc kubenswrapper[4843]: I0314 09:58:01.662976 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558038-thp2w" event={"ID":"c6fea642-5dda-4baa-8b0e-9fe644a58a2f","Type":"ContainerStarted","Data":"01891ce97e93d123cae587da704132a915b24d2b644dd245125dfc6d8c945e6d"} Mar 14 09:58:02 crc kubenswrapper[4843]: I0314 09:58:02.672961 4843 generic.go:334] "Generic (PLEG): container finished" podID="c6fea642-5dda-4baa-8b0e-9fe644a58a2f" containerID="4bb40ffe6dc4aa8b197798ffb07aef7c6e763f70ed866d6184b6c0e5ce130008" exitCode=0 Mar 14 09:58:02 crc kubenswrapper[4843]: I0314 09:58:02.673266 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558038-thp2w" event={"ID":"c6fea642-5dda-4baa-8b0e-9fe644a58a2f","Type":"ContainerDied","Data":"4bb40ffe6dc4aa8b197798ffb07aef7c6e763f70ed866d6184b6c0e5ce130008"} Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.022320 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.168920 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p985w\" (UniqueName: \"kubernetes.io/projected/c6fea642-5dda-4baa-8b0e-9fe644a58a2f-kube-api-access-p985w\") pod \"c6fea642-5dda-4baa-8b0e-9fe644a58a2f\" (UID: \"c6fea642-5dda-4baa-8b0e-9fe644a58a2f\") " Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.174564 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6fea642-5dda-4baa-8b0e-9fe644a58a2f-kube-api-access-p985w" (OuterVolumeSpecName: "kube-api-access-p985w") pod "c6fea642-5dda-4baa-8b0e-9fe644a58a2f" (UID: "c6fea642-5dda-4baa-8b0e-9fe644a58a2f"). InnerVolumeSpecName "kube-api-access-p985w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.271743 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p985w\" (UniqueName: \"kubernetes.io/projected/c6fea642-5dda-4baa-8b0e-9fe644a58a2f-kube-api-access-p985w\") on node \"crc\" DevicePath \"\"" Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.690566 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558038-thp2w" event={"ID":"c6fea642-5dda-4baa-8b0e-9fe644a58a2f","Type":"ContainerDied","Data":"01891ce97e93d123cae587da704132a915b24d2b644dd245125dfc6d8c945e6d"} Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.690821 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01891ce97e93d123cae587da704132a915b24d2b644dd245125dfc6d8c945e6d" Mar 14 09:58:04 crc kubenswrapper[4843]: I0314 09:58:04.690644 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558038-thp2w" Mar 14 09:58:05 crc kubenswrapper[4843]: I0314 09:58:05.107638 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558032-ms2dx"] Mar 14 09:58:05 crc kubenswrapper[4843]: I0314 09:58:05.124494 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558032-ms2dx"] Mar 14 09:58:05 crc kubenswrapper[4843]: I0314 09:58:05.349127 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33f82e98-788c-4584-bc61-3ed2704f00df" path="/var/lib/kubelet/pods/33f82e98-788c-4584-bc61-3ed2704f00df/volumes" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.298750 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/util/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.546433 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/pull/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.568082 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/util/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.582585 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/pull/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.741094 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/util/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.760544 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/pull/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.808036 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65270e5b79eb73ea65900207edd02f6e862f5863f7b0ae98982f030d4cpkw4d_8cd388c8-6f9c-469d-b08f-32ad35850e12/extract/0.log" Mar 14 09:58:08 crc kubenswrapper[4843]: I0314 09:58:08.933554 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/util/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.322167 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/pull/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.373764 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/util/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.468636 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/pull/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.582703 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/pull/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.603068 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/util/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.665133 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b2fc087f8a76dbce8476595fc384d33288943d0097430396a1362c765eshgz7_9892ab03-eb13-4560-a688-5e38fd015478/extract/0.log" Mar 14 09:58:09 crc kubenswrapper[4843]: I0314 09:58:09.838169 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-d47688694-ktr4z_09b1d07d-69b7-427b-9577-2a782637a2b1/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.148349 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-mxn5j_0e6cdd9e-39ed-4359-9b74-54e23a0c70e2/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.224342 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-g29z7_354f7562-fd32-491f-a3dd-b42bcf7d30cb/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.434033 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-czghc_7d4e28ff-8c1b-4533-a733-97c44a2b7b0a/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.607261 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-bmb27_3da8acd2-64eb-4f3f-b28e-80d5bb4e4743/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.640364 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-75kt6_0fbf4f22-f279-4841-b829-6d35d2dd0bfc/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.841984 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54dc5b8f8d-ffxmw_46ec1c5f-0334-4d5d-b16a-8e01ffc76613/manager/0.log" Mar 14 09:58:10 crc kubenswrapper[4843]: I0314 09:58:10.888063 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bc894d9b-8wvhp_47f2d7ba-aac2-44f4-9146-ca20192a08ab/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.126195 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7v8h2_d4d69189-dfff-4f22-b978-4988c505df47/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.176528 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-57b484b4df-fqfbh_720b9936-b2c8-4b7b-ab8f-c8ecfda9052d/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.347895 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5b6b6b4c9f-pdpx8_ddd8e1af-7084-4685-96a9-ca4eaa9a4742/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.378120 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-9dlkn_28f29a72-e196-4962-87b3-cc49b8591cc0/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.547721 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7f84474648-g59td_abe14da3-e8a8-45f6-aebb-5df03c384a35/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.650094 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-zrfcm_72b0bab9-36e7-49da-af7b-ff4fe9475545/manager/0.log" Mar 14 09:58:11 crc kubenswrapper[4843]: I0314 09:58:11.791020 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7slhjt_f5a4b439-9571-4fdb-8796-3bd729646aa3/manager/0.log" Mar 14 09:58:12 crc kubenswrapper[4843]: I0314 09:58:12.038025 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-6s6vv_8f507af6-440f-42c3-839b-0466d546a5a0/registry-server/0.log" Mar 14 09:58:12 crc kubenswrapper[4843]: I0314 09:58:12.255155 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-xctxr_92eda9b1-e96f-4a71-93f8-a595c6f5117d/manager/0.log" Mar 14 09:58:12 crc kubenswrapper[4843]: I0314 09:58:12.274486 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-59cdb7596d-l4ggb_a0339dbb-a75e-465b-a30a-75c3091f506c/manager/0.log" Mar 14 09:58:12 crc kubenswrapper[4843]: I0314 09:58:12.389812 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-tx8dt_bdefcc7b-402d-4c04-bc48-aa8b918d1b15/manager/0.log" Mar 14 09:58:12 crc kubenswrapper[4843]: I0314 09:58:12.473942 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qjs78_f75cfc62-a50d-4bad-9868-c58dc328cdbb/operator/0.log" Mar 14 09:58:12 crc kubenswrapper[4843]: I0314 09:58:12.623679 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7f9cc5dd44-kbcdm_cf4074be-d697-47bf-a44c-57822afebb15/manager/0.log" Mar 14 09:58:13 crc kubenswrapper[4843]: I0314 09:58:13.093665 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-9nkkk_99c2bbc7-1885-4031-9478-a839c36bfbc5/manager/0.log" Mar 14 09:58:13 crc kubenswrapper[4843]: I0314 09:58:13.272767 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6854b8b9d9-vgstg_f1e0e04a-1194-4fa6-878f-318d360845d3/manager/0.log" Mar 14 09:58:13 crc kubenswrapper[4843]: I0314 09:58:13.380483 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-index-7zk59_fa363db9-1707-497d-afab-dcf88242cdc2/registry-server/0.log" Mar 14 09:58:13 crc kubenswrapper[4843]: I0314 09:58:13.524738 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-66c89757b4-kgcsg_8b453c13-357e-4996-95ba-02e6d834a5b8/manager/0.log" Mar 14 09:58:15 crc kubenswrapper[4843]: I0314 09:58:15.386799 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:58:15 crc kubenswrapper[4843]: I0314 09:58:15.387039 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:58:34 crc kubenswrapper[4843]: I0314 09:58:34.509670 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-cxr6c_82f06b03-5a8e-4ef6-aac7-3c0b9326183e/control-plane-machine-set-operator/0.log" Mar 14 09:58:34 crc kubenswrapper[4843]: I0314 09:58:34.748119 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5kqvs_ce314766-eb86-4b0a-909b-2ae24850a4af/machine-api-operator/0.log" Mar 14 09:58:34 crc kubenswrapper[4843]: I0314 09:58:34.752896 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5kqvs_ce314766-eb86-4b0a-909b-2ae24850a4af/kube-rbac-proxy/0.log" Mar 14 09:58:45 crc kubenswrapper[4843]: I0314 09:58:45.387649 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:58:45 crc kubenswrapper[4843]: I0314 09:58:45.388242 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:58:49 crc kubenswrapper[4843]: I0314 09:58:48.999816 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-d6lvd_94bce3e7-2401-4f5c-919f-6cadeb6472f4/cert-manager-controller/0.log" Mar 14 09:58:49 crc kubenswrapper[4843]: I0314 09:58:49.256637 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-r4hhn_3520f5cf-b66e-4922-b114-2e7748a9103e/cert-manager-cainjector/0.log" Mar 14 09:58:49 crc kubenswrapper[4843]: I0314 09:58:49.317376 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-2rbp4_9b200cf6-c7bf-4503-9694-fb0f63d901cc/cert-manager-webhook/0.log" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.788954 4843 scope.go:117] "RemoveContainer" containerID="9b7aec10976a4de50e4011bf6809bc17b0b67d0a93693056f901317ec71a232f" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.811236 4843 scope.go:117] "RemoveContainer" containerID="9c5ec8b64bee09261423f38bef854d84055bbef2c07bcb65e7c3297828dd988b" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.847115 4843 scope.go:117] "RemoveContainer" containerID="df704e57298033b22df6bfb1642aec244e6d65c286f5e7e9971184d849a3cf49" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.912513 4843 scope.go:117] "RemoveContainer" containerID="11061fdb195c33c214175c1687a71b05388167237fb3c60cb085c9a78c6ca148" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.928236 4843 scope.go:117] "RemoveContainer" containerID="a462e3fb2aaec7845fb0dd1d1c07961c0e25ddaffc2ef3fde58df6337435f10d" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.964113 4843 scope.go:117] "RemoveContainer" containerID="c7d3d78ca542af44a37d98eda1a591d5334e6a5633feff7428bd668f99e0354b" Mar 14 09:58:56 crc kubenswrapper[4843]: I0314 09:58:56.995042 4843 scope.go:117] "RemoveContainer" containerID="3d6c1c7d106d7cfdf995dae9c299e63b33e1be9f07850d6d6ae1a85d4c9b2a3d" Mar 14 09:58:57 crc kubenswrapper[4843]: I0314 09:58:57.022504 4843 scope.go:117] "RemoveContainer" containerID="59e7029d842464cc8f2d184fae3003c0c2d520b17db970dc9b61f9f6949f03f2" Mar 14 09:58:57 crc kubenswrapper[4843]: I0314 09:58:57.067383 4843 scope.go:117] "RemoveContainer" containerID="937a6ad5ea002abf3145c18da1261d4469f21d3b2468754ae71719789c6ed630" Mar 14 09:59:04 crc kubenswrapper[4843]: I0314 09:59:04.137966 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-wwth2_22c4bb4f-2852-4c58-a0db-17793ec80671/nmstate-console-plugin/0.log" Mar 14 09:59:04 crc kubenswrapper[4843]: I0314 09:59:04.393708 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fqs5c_951a69ab-6e2a-4ff3-bcf3-db69fbfcb349/nmstate-handler/0.log" Mar 14 09:59:04 crc kubenswrapper[4843]: I0314 09:59:04.467950 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-bcwrg_c108acb5-004b-4c4e-b8e9-44aed45a2b86/kube-rbac-proxy/0.log" Mar 14 09:59:04 crc kubenswrapper[4843]: I0314 09:59:04.590561 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-bcwrg_c108acb5-004b-4c4e-b8e9-44aed45a2b86/nmstate-metrics/0.log" Mar 14 09:59:04 crc kubenswrapper[4843]: I0314 09:59:04.662781 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-n7k5d_63096d1d-7063-4821-b735-e3e65e2b9f64/nmstate-operator/0.log" Mar 14 09:59:04 crc kubenswrapper[4843]: I0314 09:59:04.765904 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-hcdc6_185eedfb-940a-4fb8-943f-3f2761f44301/nmstate-webhook/0.log" Mar 14 09:59:15 crc kubenswrapper[4843]: I0314 09:59:15.386971 4843 patch_prober.go:28] interesting pod/machine-config-daemon-gwd22 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 14 09:59:15 crc kubenswrapper[4843]: I0314 09:59:15.387784 4843 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 14 09:59:15 crc kubenswrapper[4843]: I0314 09:59:15.387873 4843 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" Mar 14 09:59:15 crc kubenswrapper[4843]: I0314 09:59:15.390733 4843 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6"} pod="openshift-machine-config-operator/machine-config-daemon-gwd22" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 14 09:59:15 crc kubenswrapper[4843]: I0314 09:59:15.390851 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerName="machine-config-daemon" containerID="cri-o://7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" gracePeriod=600 Mar 14 09:59:15 crc kubenswrapper[4843]: E0314 09:59:15.519021 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:59:16 crc kubenswrapper[4843]: I0314 09:59:16.264727 4843 generic.go:334] "Generic (PLEG): container finished" podID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" exitCode=0 Mar 14 09:59:16 crc kubenswrapper[4843]: I0314 09:59:16.264797 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerDied","Data":"7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6"} Mar 14 09:59:16 crc kubenswrapper[4843]: I0314 09:59:16.265038 4843 scope.go:117] "RemoveContainer" containerID="28e8d32901c9974c782e1e9aadfeb56ef673f0cf23ac2a95a7d8e3c463d945a9" Mar 14 09:59:16 crc kubenswrapper[4843]: I0314 09:59:16.265558 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 09:59:16 crc kubenswrapper[4843]: E0314 09:59:16.265852 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:59:20 crc kubenswrapper[4843]: I0314 09:59:20.396739 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-7vsx5_7f706824-088d-4d70-830e-4844447b900d/prometheus-operator/0.log" Mar 14 09:59:20 crc kubenswrapper[4843]: I0314 09:59:20.617949 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_7d01a80d-2445-472e-ab28-81703fb66a27/prometheus-operator-admission-webhook/0.log" Mar 14 09:59:20 crc kubenswrapper[4843]: I0314 09:59:20.902054 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_35d96605-ef77-45e3-b12c-26c192a2dd50/prometheus-operator-admission-webhook/0.log" Mar 14 09:59:21 crc kubenswrapper[4843]: I0314 09:59:21.056780 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-7l9bs_d8bb5f31-a201-4fa9-a297-f294c2aefe07/operator/0.log" Mar 14 09:59:21 crc kubenswrapper[4843]: I0314 09:59:21.124656 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-66cbf594b5-k8bb7_2c318b20-fdd8-4ebb-817b-14f94d14580c/observability-ui-dashboards/0.log" Mar 14 09:59:21 crc kubenswrapper[4843]: I0314 09:59:21.264910 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-8glvj_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c/perses-operator/0.log" Mar 14 09:59:29 crc kubenswrapper[4843]: I0314 09:59:29.339033 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 09:59:29 crc kubenswrapper[4843]: E0314 09:59:29.339686 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:59:37 crc kubenswrapper[4843]: I0314 09:59:37.619078 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-8v9j7_64b1131f-66f6-4bdf-b60e-f5d8f8680eeb/kube-rbac-proxy/0.log" Mar 14 09:59:37 crc kubenswrapper[4843]: I0314 09:59:37.705859 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-8v9j7_64b1131f-66f6-4bdf-b60e-f5d8f8680eeb/controller/0.log" Mar 14 09:59:37 crc kubenswrapper[4843]: I0314 09:59:37.807687 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-frr-files/0.log" Mar 14 09:59:37 crc kubenswrapper[4843]: I0314 09:59:37.940629 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-frr-files/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.003786 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-reloader/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.006921 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-metrics/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.037245 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-reloader/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.169146 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-frr-files/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.169264 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-reloader/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.213403 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-metrics/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.234444 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-metrics/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.415074 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-frr-files/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.425839 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-reloader/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.454966 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/cp-metrics/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.459952 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/controller/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.623820 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/frr-metrics/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.710637 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/kube-rbac-proxy/0.log" Mar 14 09:59:38 crc kubenswrapper[4843]: I0314 09:59:38.724168 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/kube-rbac-proxy-frr/0.log" Mar 14 09:59:39 crc kubenswrapper[4843]: I0314 09:59:39.021922 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/reloader/0.log" Mar 14 09:59:39 crc kubenswrapper[4843]: I0314 09:59:39.033330 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-nbb5f_2022ba85-6e88-485c-ba26-7f16fc03c257/frr-k8s-webhook-server/0.log" Mar 14 09:59:39 crc kubenswrapper[4843]: I0314 09:59:39.348088 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6c5855cb99-gcpx9_1501c094-2079-46b0-8096-b4a09ee21e8b/manager/0.log" Mar 14 09:59:39 crc kubenswrapper[4843]: I0314 09:59:39.718389 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bv84h_d3444025-ba10-41ec-94cd-0ce213c73311/kube-rbac-proxy/0.log" Mar 14 09:59:39 crc kubenswrapper[4843]: I0314 09:59:39.732787 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-76b978c99b-flxj6_ebaf629f-8937-4e21-bedb-0d5992f22edc/webhook-server/0.log" Mar 14 09:59:39 crc kubenswrapper[4843]: I0314 09:59:39.987165 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2llz5_63a81c2e-7222-4852-b557-f6a24194d46f/frr/0.log" Mar 14 09:59:40 crc kubenswrapper[4843]: I0314 09:59:40.099047 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bv84h_d3444025-ba10-41ec-94cd-0ce213c73311/speaker/0.log" Mar 14 09:59:42 crc kubenswrapper[4843]: I0314 09:59:42.338530 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 09:59:42 crc kubenswrapper[4843]: E0314 09:59:42.338836 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:59:55 crc kubenswrapper[4843]: I0314 09:59:55.339712 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 09:59:55 crc kubenswrapper[4843]: E0314 09:59:55.341037 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.226166 4843 scope.go:117] "RemoveContainer" containerID="22ee577c0712d7b442ae0692d0bed7d005fd215b2038b7fb1ad7354aaeba1bcd" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.251536 4843 scope.go:117] "RemoveContainer" containerID="c7fa90f46e3c117b478bc5938bf0826a9a90208f63cdab683cb88b9a8ee6b649" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.305345 4843 scope.go:117] "RemoveContainer" containerID="fad3d5e4665168053df9e867b63377967b5b1a4e96396a3e1e54aac815fca711" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.395157 4843 scope.go:117] "RemoveContainer" containerID="bb88982aee7a576286d16593e2eda1f05ca76fe570bdc0aa6315c4d00af5508d" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.479692 4843 scope.go:117] "RemoveContainer" containerID="87a3b4b0238d56ed494a207a396ee7efb484b018287a60013c412c82c98f0172" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.540089 4843 scope.go:117] "RemoveContainer" containerID="8e1c164eba9a345cd73527bb68de74431c2d41d2aa0373385ab4bd32751b3bb3" Mar 14 09:59:57 crc kubenswrapper[4843]: I0314 09:59:57.572340 4843 scope.go:117] "RemoveContainer" containerID="071ff12ce7fb8b9e252e7d5604e64782d0a54b63f6edec35ce091e547c16d716" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.149545 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m"] Mar 14 10:00:00 crc kubenswrapper[4843]: E0314 10:00:00.150219 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fea642-5dda-4baa-8b0e-9fe644a58a2f" containerName="oc" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.150234 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fea642-5dda-4baa-8b0e-9fe644a58a2f" containerName="oc" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.150450 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6fea642-5dda-4baa-8b0e-9fe644a58a2f" containerName="oc" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.151152 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.154877 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.155309 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.160358 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9071be7-7dbc-4e12-bb11-1b6228c57534-config-volume\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.160662 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9071be7-7dbc-4e12-bb11-1b6228c57534-secret-volume\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.160818 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62mn7\" (UniqueName: \"kubernetes.io/projected/f9071be7-7dbc-4e12-bb11-1b6228c57534-kube-api-access-62mn7\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.163500 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m"] Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.251370 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558040-vkx4b"] Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.252691 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.254860 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.255352 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.255523 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.261013 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558040-vkx4b"] Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.262106 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frcll\" (UniqueName: \"kubernetes.io/projected/0759f8a2-a59e-4675-b1ac-75ba086b6364-kube-api-access-frcll\") pod \"auto-csr-approver-29558040-vkx4b\" (UID: \"0759f8a2-a59e-4675-b1ac-75ba086b6364\") " pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.262188 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9071be7-7dbc-4e12-bb11-1b6228c57534-secret-volume\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.262254 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62mn7\" (UniqueName: \"kubernetes.io/projected/f9071be7-7dbc-4e12-bb11-1b6228c57534-kube-api-access-62mn7\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.262326 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9071be7-7dbc-4e12-bb11-1b6228c57534-config-volume\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.263669 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9071be7-7dbc-4e12-bb11-1b6228c57534-config-volume\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.269483 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9071be7-7dbc-4e12-bb11-1b6228c57534-secret-volume\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.298978 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62mn7\" (UniqueName: \"kubernetes.io/projected/f9071be7-7dbc-4e12-bb11-1b6228c57534-kube-api-access-62mn7\") pod \"collect-profiles-29558040-7nf9m\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.364419 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frcll\" (UniqueName: \"kubernetes.io/projected/0759f8a2-a59e-4675-b1ac-75ba086b6364-kube-api-access-frcll\") pod \"auto-csr-approver-29558040-vkx4b\" (UID: \"0759f8a2-a59e-4675-b1ac-75ba086b6364\") " pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.387418 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frcll\" (UniqueName: \"kubernetes.io/projected/0759f8a2-a59e-4675-b1ac-75ba086b6364-kube-api-access-frcll\") pod \"auto-csr-approver-29558040-vkx4b\" (UID: \"0759f8a2-a59e-4675-b1ac-75ba086b6364\") " pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.474296 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.568110 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:00 crc kubenswrapper[4843]: I0314 10:00:00.955886 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m"] Mar 14 10:00:01 crc kubenswrapper[4843]: I0314 10:00:01.045809 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558040-vkx4b"] Mar 14 10:00:01 crc kubenswrapper[4843]: W0314 10:00:01.058798 4843 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0759f8a2_a59e_4675_b1ac_75ba086b6364.slice/crio-180b3879f4d8e3019c6a85163cc3897a99c6f3c17ee88da5174f233506079728 WatchSource:0}: Error finding container 180b3879f4d8e3019c6a85163cc3897a99c6f3c17ee88da5174f233506079728: Status 404 returned error can't find the container with id 180b3879f4d8e3019c6a85163cc3897a99c6f3c17ee88da5174f233506079728 Mar 14 10:00:01 crc kubenswrapper[4843]: I0314 10:00:01.678157 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" event={"ID":"0759f8a2-a59e-4675-b1ac-75ba086b6364","Type":"ContainerStarted","Data":"180b3879f4d8e3019c6a85163cc3897a99c6f3c17ee88da5174f233506079728"} Mar 14 10:00:01 crc kubenswrapper[4843]: I0314 10:00:01.679997 4843 generic.go:334] "Generic (PLEG): container finished" podID="f9071be7-7dbc-4e12-bb11-1b6228c57534" containerID="a355ff4b7128a38b581c54e30393d18a307e41f26f5fca3b2511698f51476564" exitCode=0 Mar 14 10:00:01 crc kubenswrapper[4843]: I0314 10:00:01.680024 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" event={"ID":"f9071be7-7dbc-4e12-bb11-1b6228c57534","Type":"ContainerDied","Data":"a355ff4b7128a38b581c54e30393d18a307e41f26f5fca3b2511698f51476564"} Mar 14 10:00:01 crc kubenswrapper[4843]: I0314 10:00:01.680040 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" event={"ID":"f9071be7-7dbc-4e12-bb11-1b6228c57534","Type":"ContainerStarted","Data":"9f11a1a105af16dbc94da0a016617302348438643f54431992fda67137262dce"} Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.014651 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.050184 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-gw9wj"] Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.092872 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["watcher-kuttl-default/keystone-bootstrap-gw9wj"] Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.107194 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9071be7-7dbc-4e12-bb11-1b6228c57534-secret-volume\") pod \"f9071be7-7dbc-4e12-bb11-1b6228c57534\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.107324 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9071be7-7dbc-4e12-bb11-1b6228c57534-config-volume\") pod \"f9071be7-7dbc-4e12-bb11-1b6228c57534\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.107465 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62mn7\" (UniqueName: \"kubernetes.io/projected/f9071be7-7dbc-4e12-bb11-1b6228c57534-kube-api-access-62mn7\") pod \"f9071be7-7dbc-4e12-bb11-1b6228c57534\" (UID: \"f9071be7-7dbc-4e12-bb11-1b6228c57534\") " Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.109977 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9071be7-7dbc-4e12-bb11-1b6228c57534-config-volume" (OuterVolumeSpecName: "config-volume") pod "f9071be7-7dbc-4e12-bb11-1b6228c57534" (UID: "f9071be7-7dbc-4e12-bb11-1b6228c57534"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.117400 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9071be7-7dbc-4e12-bb11-1b6228c57534-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f9071be7-7dbc-4e12-bb11-1b6228c57534" (UID: "f9071be7-7dbc-4e12-bb11-1b6228c57534"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.117482 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9071be7-7dbc-4e12-bb11-1b6228c57534-kube-api-access-62mn7" (OuterVolumeSpecName: "kube-api-access-62mn7") pod "f9071be7-7dbc-4e12-bb11-1b6228c57534" (UID: "f9071be7-7dbc-4e12-bb11-1b6228c57534"). InnerVolumeSpecName "kube-api-access-62mn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.209555 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62mn7\" (UniqueName: \"kubernetes.io/projected/f9071be7-7dbc-4e12-bb11-1b6228c57534-kube-api-access-62mn7\") on node \"crc\" DevicePath \"\"" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.209588 4843 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9071be7-7dbc-4e12-bb11-1b6228c57534-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.209597 4843 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9071be7-7dbc-4e12-bb11-1b6228c57534-config-volume\") on node \"crc\" DevicePath \"\"" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.347894 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffab3465-469a-45ae-8633-0d22589ae869" path="/var/lib/kubelet/pods/ffab3465-469a-45ae-8633-0d22589ae869/volumes" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.697390 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" event={"ID":"f9071be7-7dbc-4e12-bb11-1b6228c57534","Type":"ContainerDied","Data":"9f11a1a105af16dbc94da0a016617302348438643f54431992fda67137262dce"} Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.697433 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f11a1a105af16dbc94da0a016617302348438643f54431992fda67137262dce" Mar 14 10:00:03 crc kubenswrapper[4843]: I0314 10:00:03.697477 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558040-7nf9m" Mar 14 10:00:04 crc kubenswrapper[4843]: I0314 10:00:04.078509 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz"] Mar 14 10:00:04 crc kubenswrapper[4843]: I0314 10:00:04.088284 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557995-sd7tz"] Mar 14 10:00:04 crc kubenswrapper[4843]: I0314 10:00:04.575894 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_120dfd6f-cc76-42f9-83fb-954ae00d21c4/init-config-reloader/0.log" Mar 14 10:00:04 crc kubenswrapper[4843]: I0314 10:00:04.854268 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_120dfd6f-cc76-42f9-83fb-954ae00d21c4/init-config-reloader/0.log" Mar 14 10:00:04 crc kubenswrapper[4843]: I0314 10:00:04.896165 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_120dfd6f-cc76-42f9-83fb-954ae00d21c4/alertmanager/0.log" Mar 14 10:00:04 crc kubenswrapper[4843]: I0314 10:00:04.907900 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_alertmanager-metric-storage-0_120dfd6f-cc76-42f9-83fb-954ae00d21c4/config-reloader/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.139979 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_0b50529a-c274-4664-b0ef-a06045bbfb96/proxy-httpd/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.151071 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_0b50529a-c274-4664-b0ef-a06045bbfb96/ceilometer-central-agent/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.170939 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_0b50529a-c274-4664-b0ef-a06045bbfb96/ceilometer-notification-agent/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.253903 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_ceilometer-0_0b50529a-c274-4664-b0ef-a06045bbfb96/sg-core/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.348815 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5627cd7c-aab3-40a1-b844-9c4d5fba908f" path="/var/lib/kubelet/pods/5627cd7c-aab3-40a1-b844-9c4d5fba908f/volumes" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.468096 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_kube-state-metrics-0_4e9ebb9e-ebf4-4700-b5e2-cd76d5bccced/kube-state-metrics/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.490728 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_keystone-7f5697c5fb-hqksz_494a65cb-9ba0-448d-944c-7040f8e30bf7/keystone-api/0.log" Mar 14 10:00:05 crc kubenswrapper[4843]: I0314 10:00:05.811462 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstack-galera-0_fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43/mysql-bootstrap/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.045940 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstack-galera-0_fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43/galera/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.046233 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstack-galera-0_fe7fb9b9-cb1c-47d3-be1e-461c7ea83b43/mysql-bootstrap/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.446698 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_openstackclient_c3ef3e95-e2e0-456f-9262-2ae37fb42668/openstackclient/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.606566 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2/init-config-reloader/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.723676 4843 generic.go:334] "Generic (PLEG): container finished" podID="0759f8a2-a59e-4675-b1ac-75ba086b6364" containerID="030417ebba033b3870b459c003a78b35378e62bb638a21d1f64b6dc8d7a6fff0" exitCode=0 Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.723714 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" event={"ID":"0759f8a2-a59e-4675-b1ac-75ba086b6364","Type":"ContainerDied","Data":"030417ebba033b3870b459c003a78b35378e62bb638a21d1f64b6dc8d7a6fff0"} Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.770087 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2/config-reloader/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.801470 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2/init-config-reloader/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.897441 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2/prometheus/0.log" Mar 14 10:00:06 crc kubenswrapper[4843]: I0314 10:00:06.976661 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_prometheus-metric-storage-0_b4349fd9-0816-49bf-9c0c-c3cbbabe2ff2/thanos-sidecar/0.log" Mar 14 10:00:07 crc kubenswrapper[4843]: I0314 10:00:07.108144 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-notifications-server-0_eaec015a-59c1-492c-9487-003e222f49fc/setup-container/0.log" Mar 14 10:00:07 crc kubenswrapper[4843]: I0314 10:00:07.358950 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-notifications-server-0_eaec015a-59c1-492c-9487-003e222f49fc/setup-container/0.log" Mar 14 10:00:07 crc kubenswrapper[4843]: I0314 10:00:07.394003 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-notifications-server-0_eaec015a-59c1-492c-9487-003e222f49fc/rabbitmq/0.log" Mar 14 10:00:07 crc kubenswrapper[4843]: I0314 10:00:07.562933 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-server-0_ced48f61-fae8-48aa-819f-c26f7296dac5/setup-container/0.log" Mar 14 10:00:07 crc kubenswrapper[4843]: I0314 10:00:07.903903 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-server-0_ced48f61-fae8-48aa-819f-c26f7296dac5/setup-container/0.log" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.043078 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.063563 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_rabbitmq-server-0_ced48f61-fae8-48aa-819f-c26f7296dac5/rabbitmq/0.log" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.190769 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frcll\" (UniqueName: \"kubernetes.io/projected/0759f8a2-a59e-4675-b1ac-75ba086b6364-kube-api-access-frcll\") pod \"0759f8a2-a59e-4675-b1ac-75ba086b6364\" (UID: \"0759f8a2-a59e-4675-b1ac-75ba086b6364\") " Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.197089 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0759f8a2-a59e-4675-b1ac-75ba086b6364-kube-api-access-frcll" (OuterVolumeSpecName: "kube-api-access-frcll") pod "0759f8a2-a59e-4675-b1ac-75ba086b6364" (UID: "0759f8a2-a59e-4675-b1ac-75ba086b6364"). InnerVolumeSpecName "kube-api-access-frcll". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.292194 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frcll\" (UniqueName: \"kubernetes.io/projected/0759f8a2-a59e-4675-b1ac-75ba086b6364-kube-api-access-frcll\") on node \"crc\" DevicePath \"\"" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.338380 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:00:08 crc kubenswrapper[4843]: E0314 10:00:08.338675 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.740191 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" event={"ID":"0759f8a2-a59e-4675-b1ac-75ba086b6364","Type":"ContainerDied","Data":"180b3879f4d8e3019c6a85163cc3897a99c6f3c17ee88da5174f233506079728"} Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.740231 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="180b3879f4d8e3019c6a85163cc3897a99c6f3c17ee88da5174f233506079728" Mar 14 10:00:08 crc kubenswrapper[4843]: I0314 10:00:08.740301 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558040-vkx4b" Mar 14 10:00:09 crc kubenswrapper[4843]: I0314 10:00:09.107929 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558034-cf7wr"] Mar 14 10:00:09 crc kubenswrapper[4843]: I0314 10:00:09.115765 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558034-cf7wr"] Mar 14 10:00:09 crc kubenswrapper[4843]: I0314 10:00:09.347730 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="899a1e8a-4cee-4187-9d37-9afcc5c310db" path="/var/lib/kubelet/pods/899a1e8a-4cee-4187-9d37-9afcc5c310db/volumes" Mar 14 10:00:15 crc kubenswrapper[4843]: I0314 10:00:15.808679 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/watcher-kuttl-default_memcached-0_08eba4dc-5f7e-47fe-994d-72af55c68a77/memcached/0.log" Mar 14 10:00:21 crc kubenswrapper[4843]: I0314 10:00:21.338832 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:00:21 crc kubenswrapper[4843]: E0314 10:00:21.339503 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:00:26 crc kubenswrapper[4843]: I0314 10:00:26.680655 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/util/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.028338 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/pull/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.075945 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/util/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.081394 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/pull/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.262036 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/pull/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.269175 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/util/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.344299 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cs2f2_e952e216-1adb-456a-9f7e-61d2dc2ddb78/extract/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.545950 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/util/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.701614 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/pull/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.755661 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/util/0.log" Mar 14 10:00:27 crc kubenswrapper[4843]: I0314 10:00:27.785360 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/pull/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.000571 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/pull/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.014201 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/util/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.034892 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c14m429_d916f254-7cc8-48b9-b01e-ba6270584eb3/extract/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.220966 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/util/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.466550 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/util/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.500706 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/pull/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.530391 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/pull/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.668406 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/util/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.677083 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/pull/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.713738 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5qhxpx_bd5a3603-199d-45f6-80c7-afb96e29068f/extract/0.log" Mar 14 10:00:28 crc kubenswrapper[4843]: I0314 10:00:28.930708 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/util/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.066063 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/pull/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.081770 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/pull/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.091640 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/util/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.263082 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/util/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.284930 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/pull/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.305773 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f0857cj7_cb06d762-fe54-4e6f-917e-1431b5948d9a/extract/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.522296 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/extract-utilities/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.678807 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/extract-content/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.682146 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/extract-utilities/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.723533 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/extract-content/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.859150 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/extract-utilities/0.log" Mar 14 10:00:29 crc kubenswrapper[4843]: I0314 10:00:29.925469 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/extract-content/0.log" Mar 14 10:00:30 crc kubenswrapper[4843]: I0314 10:00:30.106565 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/extract-utilities/0.log" Mar 14 10:00:30 crc kubenswrapper[4843]: I0314 10:00:30.606492 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tt7cz_38ba05ff-31cc-4fa9-a6fc-e374e8c6ede6/registry-server/0.log" Mar 14 10:00:30 crc kubenswrapper[4843]: I0314 10:00:30.839483 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/extract-utilities/0.log" Mar 14 10:00:30 crc kubenswrapper[4843]: I0314 10:00:30.900310 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/extract-content/0.log" Mar 14 10:00:30 crc kubenswrapper[4843]: I0314 10:00:30.903868 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/extract-content/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.068331 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/extract-content/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.130024 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/extract-utilities/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.255942 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-565tx_3bd512f7-d36b-43b0-94af-cee8982d7630/marketplace-operator/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.500802 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/extract-utilities/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.642840 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-llm5h_353b52a0-aebf-426a-9006-74c1d24e8c48/registry-server/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.769161 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/extract-utilities/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.782031 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/extract-content/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.813794 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/extract-content/0.log" Mar 14 10:00:31 crc kubenswrapper[4843]: I0314 10:00:31.964440 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/extract-content/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.035884 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/extract-utilities/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.064067 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/extract-utilities/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.132044 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tfbnl_166fa22e-f961-4d02-bd2a-a6e215204f4b/registry-server/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.251640 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/extract-content/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.289130 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/extract-utilities/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.290796 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/extract-content/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.452951 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/extract-content/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.500408 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/extract-utilities/0.log" Mar 14 10:00:32 crc kubenswrapper[4843]: I0314 10:00:32.907880 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vx4kc_f9c9423f-f5aa-4a79-8084-f775e9a8af97/registry-server/0.log" Mar 14 10:00:36 crc kubenswrapper[4843]: I0314 10:00:36.339342 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:00:36 crc kubenswrapper[4843]: E0314 10:00:36.339890 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:00:46 crc kubenswrapper[4843]: I0314 10:00:46.551119 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-69fb744b97-5knkk_7d01a80d-2445-472e-ab28-81703fb66a27/prometheus-operator-admission-webhook/0.log" Mar 14 10:00:46 crc kubenswrapper[4843]: I0314 10:00:46.562051 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-7vsx5_7f706824-088d-4d70-830e-4844447b900d/prometheus-operator/0.log" Mar 14 10:00:46 crc kubenswrapper[4843]: I0314 10:00:46.646135 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-69fb744b97-scrd4_35d96605-ef77-45e3-b12c-26c192a2dd50/prometheus-operator-admission-webhook/0.log" Mar 14 10:00:46 crc kubenswrapper[4843]: I0314 10:00:46.796021 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-66cbf594b5-k8bb7_2c318b20-fdd8-4ebb-817b-14f94d14580c/observability-ui-dashboards/0.log" Mar 14 10:00:46 crc kubenswrapper[4843]: I0314 10:00:46.847663 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-7l9bs_d8bb5f31-a201-4fa9-a297-f294c2aefe07/operator/0.log" Mar 14 10:00:46 crc kubenswrapper[4843]: I0314 10:00:46.947481 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-8glvj_5e9fbec7-6a0e-48cf-95fb-5a470a77ff4c/perses-operator/0.log" Mar 14 10:00:49 crc kubenswrapper[4843]: I0314 10:00:49.341709 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:00:49 crc kubenswrapper[4843]: E0314 10:00:49.342218 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:00:57 crc kubenswrapper[4843]: I0314 10:00:57.696541 4843 scope.go:117] "RemoveContainer" containerID="6e944e23ca4b716eb4ca56bdc9761b797b2ed9c7df579a570eed58ad6368fe95" Mar 14 10:00:57 crc kubenswrapper[4843]: I0314 10:00:57.765129 4843 scope.go:117] "RemoveContainer" containerID="f403133beb54c29f3eaf7d90bd4982ce8a220d34a36d2861d39cb44edd63d32c" Mar 14 10:00:58 crc kubenswrapper[4843]: I0314 10:00:57.795648 4843 scope.go:117] "RemoveContainer" containerID="47018f9f60d648e7cb8a81aa957e48673fee220b341a002cbd5db188bea1b9e1" Mar 14 10:00:58 crc kubenswrapper[4843]: I0314 10:00:58.509696 4843 scope.go:117] "RemoveContainer" containerID="8f8524a861ca84f25b68de2f13eebedf2cdef3501e1555f9efd60bc6a2c1c648" Mar 14 10:00:58 crc kubenswrapper[4843]: I0314 10:00:58.536686 4843 scope.go:117] "RemoveContainer" containerID="957acd9fa549ac74b16cbd40d39d256c0886dddf20782ba9cb8142060f6adc9d" Mar 14 10:00:58 crc kubenswrapper[4843]: I0314 10:00:58.555850 4843 scope.go:117] "RemoveContainer" containerID="39518908e826d8a0f2bf3e7f8ca5097f2aa3bddfc661fced13a318202bf7f790" Mar 14 10:00:58 crc kubenswrapper[4843]: I0314 10:00:58.611505 4843 scope.go:117] "RemoveContainer" containerID="94e5e55eb4e83ba29e712badf5917865efd660c2b0fd04d836960d5e61d37015" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.163652 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["watcher-kuttl-default/keystone-cron-29558041-khbnb"] Mar 14 10:01:00 crc kubenswrapper[4843]: E0314 10:01:00.164348 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9071be7-7dbc-4e12-bb11-1b6228c57534" containerName="collect-profiles" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.164367 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9071be7-7dbc-4e12-bb11-1b6228c57534" containerName="collect-profiles" Mar 14 10:01:00 crc kubenswrapper[4843]: E0314 10:01:00.164399 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0759f8a2-a59e-4675-b1ac-75ba086b6364" containerName="oc" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.164407 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="0759f8a2-a59e-4675-b1ac-75ba086b6364" containerName="oc" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.164647 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="0759f8a2-a59e-4675-b1ac-75ba086b6364" containerName="oc" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.164665 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9071be7-7dbc-4e12-bb11-1b6228c57534" containerName="collect-profiles" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.165390 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.175499 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-cron-29558041-khbnb"] Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.292973 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-config-data\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.293022 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-fernet-keys\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.293056 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-combined-ca-bundle\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.293093 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-cert-memcached-mtls\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.293242 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhvpr\" (UniqueName: \"kubernetes.io/projected/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-kube-api-access-lhvpr\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.395641 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-config-data\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.395736 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-fernet-keys\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.395804 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-combined-ca-bundle\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.395887 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-cert-memcached-mtls\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.396120 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhvpr\" (UniqueName: \"kubernetes.io/projected/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-kube-api-access-lhvpr\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.408188 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-config-data\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.408296 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-cert-memcached-mtls\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.408343 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-fernet-keys\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.412069 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-combined-ca-bundle\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.419304 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhvpr\" (UniqueName: \"kubernetes.io/projected/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-kube-api-access-lhvpr\") pod \"keystone-cron-29558041-khbnb\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.485717 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:00 crc kubenswrapper[4843]: I0314 10:01:00.940837 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["watcher-kuttl-default/keystone-cron-29558041-khbnb"] Mar 14 10:01:01 crc kubenswrapper[4843]: I0314 10:01:01.144670 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" event={"ID":"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd","Type":"ContainerStarted","Data":"024e8ea6199b48e2f38425a4dd8b43d49efa029bf399feb2e7e7f065367a10df"} Mar 14 10:01:01 crc kubenswrapper[4843]: I0314 10:01:01.145243 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" event={"ID":"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd","Type":"ContainerStarted","Data":"61626f0906dc2c2ed8f1c6d6ebdfbea6fcbe915e07749b6f4eb347dab031b2f2"} Mar 14 10:01:01 crc kubenswrapper[4843]: I0314 10:01:01.172517 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" podStartSLOduration=1.172474566 podStartE2EDuration="1.172474566s" podCreationTimestamp="2026-03-14 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-14 10:01:01.16977737 +0000 UTC m=+3028.482388498" watchObservedRunningTime="2026-03-14 10:01:01.172474566 +0000 UTC m=+3028.485085694" Mar 14 10:01:03 crc kubenswrapper[4843]: I0314 10:01:03.162887 4843 generic.go:334] "Generic (PLEG): container finished" podID="e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" containerID="024e8ea6199b48e2f38425a4dd8b43d49efa029bf399feb2e7e7f065367a10df" exitCode=0 Mar 14 10:01:03 crc kubenswrapper[4843]: I0314 10:01:03.163232 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" event={"ID":"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd","Type":"ContainerDied","Data":"024e8ea6199b48e2f38425a4dd8b43d49efa029bf399feb2e7e7f065367a10df"} Mar 14 10:01:03 crc kubenswrapper[4843]: I0314 10:01:03.348815 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:01:03 crc kubenswrapper[4843]: E0314 10:01:03.349097 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.471062 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.566263 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhvpr\" (UniqueName: \"kubernetes.io/projected/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-kube-api-access-lhvpr\") pod \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.566339 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-config-data\") pod \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.566374 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-combined-ca-bundle\") pod \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.566537 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-fernet-keys\") pod \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.566627 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-cert-memcached-mtls\") pod \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\" (UID: \"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd\") " Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.571321 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-kube-api-access-lhvpr" (OuterVolumeSpecName: "kube-api-access-lhvpr") pod "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" (UID: "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd"). InnerVolumeSpecName "kube-api-access-lhvpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.580373 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" (UID: "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.599450 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" (UID: "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.614675 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-config-data" (OuterVolumeSpecName: "config-data") pod "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" (UID: "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.630376 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-cert-memcached-mtls" (OuterVolumeSpecName: "cert-memcached-mtls") pod "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" (UID: "e95cc8ec-02fb-4344-a08e-ccd9a99f47bd"). InnerVolumeSpecName "cert-memcached-mtls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.668311 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhvpr\" (UniqueName: \"kubernetes.io/projected/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-kube-api-access-lhvpr\") on node \"crc\" DevicePath \"\"" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.668347 4843 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-config-data\") on node \"crc\" DevicePath \"\"" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.668356 4843 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.668366 4843 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 14 10:01:04 crc kubenswrapper[4843]: I0314 10:01:04.668375 4843 reconciler_common.go:293] "Volume detached for volume \"cert-memcached-mtls\" (UniqueName: \"kubernetes.io/secret/e95cc8ec-02fb-4344-a08e-ccd9a99f47bd-cert-memcached-mtls\") on node \"crc\" DevicePath \"\"" Mar 14 10:01:05 crc kubenswrapper[4843]: I0314 10:01:05.179856 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" event={"ID":"e95cc8ec-02fb-4344-a08e-ccd9a99f47bd","Type":"ContainerDied","Data":"61626f0906dc2c2ed8f1c6d6ebdfbea6fcbe915e07749b6f4eb347dab031b2f2"} Mar 14 10:01:05 crc kubenswrapper[4843]: I0314 10:01:05.179900 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61626f0906dc2c2ed8f1c6d6ebdfbea6fcbe915e07749b6f4eb347dab031b2f2" Mar 14 10:01:05 crc kubenswrapper[4843]: I0314 10:01:05.179899 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="watcher-kuttl-default/keystone-cron-29558041-khbnb" Mar 14 10:01:15 crc kubenswrapper[4843]: I0314 10:01:15.340022 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:01:15 crc kubenswrapper[4843]: E0314 10:01:15.340862 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:01:26 crc kubenswrapper[4843]: I0314 10:01:26.339726 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:01:26 crc kubenswrapper[4843]: E0314 10:01:26.340401 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:01:37 crc kubenswrapper[4843]: I0314 10:01:37.344755 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:01:37 crc kubenswrapper[4843]: E0314 10:01:37.345835 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:01:48 crc kubenswrapper[4843]: I0314 10:01:48.339122 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:01:48 crc kubenswrapper[4843]: E0314 10:01:48.340120 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:01:59 crc kubenswrapper[4843]: I0314 10:01:59.202972 4843 scope.go:117] "RemoveContainer" containerID="3d8f7c84a9f2927c286166a5bf0d0ea9bf0bc5825f41f68d315f57ccad05294e" Mar 14 10:01:59 crc kubenswrapper[4843]: I0314 10:01:59.267440 4843 scope.go:117] "RemoveContainer" containerID="6f0119e162497e6f567603380b44c3b32a8cf39558a29296139d301f1d7bff79" Mar 14 10:01:59 crc kubenswrapper[4843]: I0314 10:01:59.289186 4843 scope.go:117] "RemoveContainer" containerID="0fc03499f13ec2dd1367f7d421a08174ac9963e0c5df06c41a130cc4f31ed08d" Mar 14 10:01:59 crc kubenswrapper[4843]: I0314 10:01:59.325112 4843 scope.go:117] "RemoveContainer" containerID="ccf207a3097e578e80bde5360a5059a4945a7bd296718a419144f6085e14980f" Mar 14 10:01:59 crc kubenswrapper[4843]: I0314 10:01:59.338765 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:01:59 crc kubenswrapper[4843]: E0314 10:01:59.339055 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:01:59 crc kubenswrapper[4843]: E0314 10:01:59.384176 4843 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c313e64_65bc_4dfd_a1cf_61956ec5870e.slice/crio-bac7c0be5cf494e8c6e4d713e9b7c582f3278fb20ce89aae26628efc5a628cf6.scope\": RecentStats: unable to find data in memory cache]" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.143253 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558042-gbkvh"] Mar 14 10:02:00 crc kubenswrapper[4843]: E0314 10:02:00.143752 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" containerName="keystone-cron" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.143776 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" containerName="keystone-cron" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.143962 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="e95cc8ec-02fb-4344-a08e-ccd9a99f47bd" containerName="keystone-cron" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.144644 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.146958 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.147033 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.147050 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.154455 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558042-gbkvh"] Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.199193 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4zl9\" (UniqueName: \"kubernetes.io/projected/b20e688b-1831-41ce-8b34-e3af8a9fd5c5-kube-api-access-p4zl9\") pod \"auto-csr-approver-29558042-gbkvh\" (UID: \"b20e688b-1831-41ce-8b34-e3af8a9fd5c5\") " pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.216313 4843 generic.go:334] "Generic (PLEG): container finished" podID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerID="bac7c0be5cf494e8c6e4d713e9b7c582f3278fb20ce89aae26628efc5a628cf6" exitCode=0 Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.216357 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s74vk/must-gather-88lrc" event={"ID":"4c313e64-65bc-4dfd-a1cf-61956ec5870e","Type":"ContainerDied","Data":"bac7c0be5cf494e8c6e4d713e9b7c582f3278fb20ce89aae26628efc5a628cf6"} Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.216864 4843 scope.go:117] "RemoveContainer" containerID="bac7c0be5cf494e8c6e4d713e9b7c582f3278fb20ce89aae26628efc5a628cf6" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.301108 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4zl9\" (UniqueName: \"kubernetes.io/projected/b20e688b-1831-41ce-8b34-e3af8a9fd5c5-kube-api-access-p4zl9\") pod \"auto-csr-approver-29558042-gbkvh\" (UID: \"b20e688b-1831-41ce-8b34-e3af8a9fd5c5\") " pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.330659 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4zl9\" (UniqueName: \"kubernetes.io/projected/b20e688b-1831-41ce-8b34-e3af8a9fd5c5-kube-api-access-p4zl9\") pod \"auto-csr-approver-29558042-gbkvh\" (UID: \"b20e688b-1831-41ce-8b34-e3af8a9fd5c5\") " pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.382758 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s74vk_must-gather-88lrc_4c313e64-65bc-4dfd-a1cf-61956ec5870e/gather/0.log" Mar 14 10:02:00 crc kubenswrapper[4843]: I0314 10:02:00.481992 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:01 crc kubenswrapper[4843]: I0314 10:02:01.004468 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558042-gbkvh"] Mar 14 10:02:01 crc kubenswrapper[4843]: I0314 10:02:01.032933 4843 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 14 10:02:01 crc kubenswrapper[4843]: I0314 10:02:01.244134 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" event={"ID":"b20e688b-1831-41ce-8b34-e3af8a9fd5c5","Type":"ContainerStarted","Data":"8558d5c06c957bd28f66a22207daf661a0a69cf61b03b15e5c410583748dd12a"} Mar 14 10:02:02 crc kubenswrapper[4843]: I0314 10:02:02.254368 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" event={"ID":"b20e688b-1831-41ce-8b34-e3af8a9fd5c5","Type":"ContainerStarted","Data":"6a3b7cbcb09dcd774da77cc93e9995b85054c42ce8b6c345d31d18e3a62b3a6c"} Mar 14 10:02:03 crc kubenswrapper[4843]: I0314 10:02:03.261624 4843 generic.go:334] "Generic (PLEG): container finished" podID="b20e688b-1831-41ce-8b34-e3af8a9fd5c5" containerID="6a3b7cbcb09dcd774da77cc93e9995b85054c42ce8b6c345d31d18e3a62b3a6c" exitCode=0 Mar 14 10:02:03 crc kubenswrapper[4843]: I0314 10:02:03.261672 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" event={"ID":"b20e688b-1831-41ce-8b34-e3af8a9fd5c5","Type":"ContainerDied","Data":"6a3b7cbcb09dcd774da77cc93e9995b85054c42ce8b6c345d31d18e3a62b3a6c"} Mar 14 10:02:04 crc kubenswrapper[4843]: I0314 10:02:04.653357 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:04 crc kubenswrapper[4843]: I0314 10:02:04.691895 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4zl9\" (UniqueName: \"kubernetes.io/projected/b20e688b-1831-41ce-8b34-e3af8a9fd5c5-kube-api-access-p4zl9\") pod \"b20e688b-1831-41ce-8b34-e3af8a9fd5c5\" (UID: \"b20e688b-1831-41ce-8b34-e3af8a9fd5c5\") " Mar 14 10:02:04 crc kubenswrapper[4843]: I0314 10:02:04.709976 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b20e688b-1831-41ce-8b34-e3af8a9fd5c5-kube-api-access-p4zl9" (OuterVolumeSpecName: "kube-api-access-p4zl9") pod "b20e688b-1831-41ce-8b34-e3af8a9fd5c5" (UID: "b20e688b-1831-41ce-8b34-e3af8a9fd5c5"). InnerVolumeSpecName "kube-api-access-p4zl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 10:02:04 crc kubenswrapper[4843]: I0314 10:02:04.793434 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4zl9\" (UniqueName: \"kubernetes.io/projected/b20e688b-1831-41ce-8b34-e3af8a9fd5c5-kube-api-access-p4zl9\") on node \"crc\" DevicePath \"\"" Mar 14 10:02:05 crc kubenswrapper[4843]: I0314 10:02:05.277523 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" event={"ID":"b20e688b-1831-41ce-8b34-e3af8a9fd5c5","Type":"ContainerDied","Data":"8558d5c06c957bd28f66a22207daf661a0a69cf61b03b15e5c410583748dd12a"} Mar 14 10:02:05 crc kubenswrapper[4843]: I0314 10:02:05.277562 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8558d5c06c957bd28f66a22207daf661a0a69cf61b03b15e5c410583748dd12a" Mar 14 10:02:05 crc kubenswrapper[4843]: I0314 10:02:05.277590 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558042-gbkvh" Mar 14 10:02:05 crc kubenswrapper[4843]: I0314 10:02:05.327793 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558036-9tltz"] Mar 14 10:02:05 crc kubenswrapper[4843]: I0314 10:02:05.334546 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558036-9tltz"] Mar 14 10:02:05 crc kubenswrapper[4843]: I0314 10:02:05.355221 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="367c08b3-ceb2-4787-841b-c8bb6d3ce971" path="/var/lib/kubelet/pods/367c08b3-ceb2-4787-841b-c8bb6d3ce971/volumes" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.176987 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s74vk/must-gather-88lrc"] Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.177765 4843 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-s74vk/must-gather-88lrc" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="copy" containerID="cri-o://2dc2d95ed3d84c4fe62cd7638b6419f3992efdffa94360126f8f6def20ba592c" gracePeriod=2 Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.185984 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s74vk/must-gather-88lrc"] Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.302534 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s74vk_must-gather-88lrc_4c313e64-65bc-4dfd-a1cf-61956ec5870e/copy/0.log" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.304423 4843 generic.go:334] "Generic (PLEG): container finished" podID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerID="2dc2d95ed3d84c4fe62cd7638b6419f3992efdffa94360126f8f6def20ba592c" exitCode=143 Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.543226 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s74vk_must-gather-88lrc_4c313e64-65bc-4dfd-a1cf-61956ec5870e/copy/0.log" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.544148 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.650540 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4c313e64-65bc-4dfd-a1cf-61956ec5870e-must-gather-output\") pod \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.650688 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmwx4\" (UniqueName: \"kubernetes.io/projected/4c313e64-65bc-4dfd-a1cf-61956ec5870e-kube-api-access-jmwx4\") pod \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\" (UID: \"4c313e64-65bc-4dfd-a1cf-61956ec5870e\") " Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.655983 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c313e64-65bc-4dfd-a1cf-61956ec5870e-kube-api-access-jmwx4" (OuterVolumeSpecName: "kube-api-access-jmwx4") pod "4c313e64-65bc-4dfd-a1cf-61956ec5870e" (UID: "4c313e64-65bc-4dfd-a1cf-61956ec5870e"). InnerVolumeSpecName "kube-api-access-jmwx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.752604 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmwx4\" (UniqueName: \"kubernetes.io/projected/4c313e64-65bc-4dfd-a1cf-61956ec5870e-kube-api-access-jmwx4\") on node \"crc\" DevicePath \"\"" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.780758 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c313e64-65bc-4dfd-a1cf-61956ec5870e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4c313e64-65bc-4dfd-a1cf-61956ec5870e" (UID: "4c313e64-65bc-4dfd-a1cf-61956ec5870e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 14 10:02:08 crc kubenswrapper[4843]: I0314 10:02:08.853965 4843 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4c313e64-65bc-4dfd-a1cf-61956ec5870e-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 14 10:02:09 crc kubenswrapper[4843]: I0314 10:02:09.314608 4843 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s74vk_must-gather-88lrc_4c313e64-65bc-4dfd-a1cf-61956ec5870e/copy/0.log" Mar 14 10:02:09 crc kubenswrapper[4843]: I0314 10:02:09.315255 4843 scope.go:117] "RemoveContainer" containerID="2dc2d95ed3d84c4fe62cd7638b6419f3992efdffa94360126f8f6def20ba592c" Mar 14 10:02:09 crc kubenswrapper[4843]: I0314 10:02:09.315321 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s74vk/must-gather-88lrc" Mar 14 10:02:09 crc kubenswrapper[4843]: I0314 10:02:09.361384 4843 scope.go:117] "RemoveContainer" containerID="bac7c0be5cf494e8c6e4d713e9b7c582f3278fb20ce89aae26628efc5a628cf6" Mar 14 10:02:09 crc kubenswrapper[4843]: I0314 10:02:09.365668 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" path="/var/lib/kubelet/pods/4c313e64-65bc-4dfd-a1cf-61956ec5870e/volumes" Mar 14 10:02:11 crc kubenswrapper[4843]: I0314 10:02:11.339463 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:02:11 crc kubenswrapper[4843]: E0314 10:02:11.340708 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:02:22 crc kubenswrapper[4843]: I0314 10:02:22.339036 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:02:22 crc kubenswrapper[4843]: E0314 10:02:22.339800 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:02:37 crc kubenswrapper[4843]: I0314 10:02:37.339717 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:02:37 crc kubenswrapper[4843]: E0314 10:02:37.340461 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:02:52 crc kubenswrapper[4843]: I0314 10:02:52.339509 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:02:52 crc kubenswrapper[4843]: E0314 10:02:52.340927 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:02:59 crc kubenswrapper[4843]: I0314 10:02:59.435855 4843 scope.go:117] "RemoveContainer" containerID="3f37680f9b8e4cdee591aea42a93e501cdf46ecde5506a6f974159bc264de966" Mar 14 10:02:59 crc kubenswrapper[4843]: I0314 10:02:59.486646 4843 scope.go:117] "RemoveContainer" containerID="faaee28c2ddf305c67e11fee6328d5944040b414ad01ef9559666019cce0fa73" Mar 14 10:03:04 crc kubenswrapper[4843]: I0314 10:03:04.338706 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:03:04 crc kubenswrapper[4843]: E0314 10:03:04.339198 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:03:17 crc kubenswrapper[4843]: I0314 10:03:17.338723 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:03:17 crc kubenswrapper[4843]: E0314 10:03:17.339547 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:03:29 crc kubenswrapper[4843]: I0314 10:03:29.338880 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:03:29 crc kubenswrapper[4843]: E0314 10:03:29.340678 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:03:44 crc kubenswrapper[4843]: I0314 10:03:44.339316 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:03:44 crc kubenswrapper[4843]: E0314 10:03:44.340555 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:03:57 crc kubenswrapper[4843]: I0314 10:03:57.338831 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:03:57 crc kubenswrapper[4843]: E0314 10:03:57.339595 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.200136 4843 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558044-97x8m"] Mar 14 10:04:00 crc kubenswrapper[4843]: E0314 10:04:00.200802 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="copy" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.200817 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="copy" Mar 14 10:04:00 crc kubenswrapper[4843]: E0314 10:04:00.200843 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b20e688b-1831-41ce-8b34-e3af8a9fd5c5" containerName="oc" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.200851 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="b20e688b-1831-41ce-8b34-e3af8a9fd5c5" containerName="oc" Mar 14 10:04:00 crc kubenswrapper[4843]: E0314 10:04:00.200861 4843 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="gather" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.200871 4843 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="gather" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.201040 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="b20e688b-1831-41ce-8b34-e3af8a9fd5c5" containerName="oc" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.201065 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="gather" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.201078 4843 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c313e64-65bc-4dfd-a1cf-61956ec5870e" containerName="copy" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.201774 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.204864 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.205165 4843 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.205349 4843 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-xdk8j" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.210288 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558044-97x8m"] Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.294690 4843 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc59p\" (UniqueName: \"kubernetes.io/projected/52a5287e-5893-44d1-b5ab-ee535f4bec9a-kube-api-access-sc59p\") pod \"auto-csr-approver-29558044-97x8m\" (UID: \"52a5287e-5893-44d1-b5ab-ee535f4bec9a\") " pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.396337 4843 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc59p\" (UniqueName: \"kubernetes.io/projected/52a5287e-5893-44d1-b5ab-ee535f4bec9a-kube-api-access-sc59p\") pod \"auto-csr-approver-29558044-97x8m\" (UID: \"52a5287e-5893-44d1-b5ab-ee535f4bec9a\") " pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.423459 4843 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc59p\" (UniqueName: \"kubernetes.io/projected/52a5287e-5893-44d1-b5ab-ee535f4bec9a-kube-api-access-sc59p\") pod \"auto-csr-approver-29558044-97x8m\" (UID: \"52a5287e-5893-44d1-b5ab-ee535f4bec9a\") " pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.527868 4843 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:00 crc kubenswrapper[4843]: I0314 10:04:00.958046 4843 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558044-97x8m"] Mar 14 10:04:01 crc kubenswrapper[4843]: I0314 10:04:01.259937 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558044-97x8m" event={"ID":"52a5287e-5893-44d1-b5ab-ee535f4bec9a","Type":"ContainerStarted","Data":"31e6a27fd33c1170ff53bbf7cf4863af0588f9a2122fa79826f359cdbb86b720"} Mar 14 10:04:02 crc kubenswrapper[4843]: I0314 10:04:02.269261 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558044-97x8m" event={"ID":"52a5287e-5893-44d1-b5ab-ee535f4bec9a","Type":"ContainerStarted","Data":"89e033162395d140342b8dc679ed476e434560bbb88cefc21a58ee59156b4332"} Mar 14 10:04:02 crc kubenswrapper[4843]: I0314 10:04:02.287789 4843 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558044-97x8m" podStartSLOduration=1.420125297 podStartE2EDuration="2.287771091s" podCreationTimestamp="2026-03-14 10:04:00 +0000 UTC" firstStartedPulling="2026-03-14 10:04:00.96126605 +0000 UTC m=+3208.273877178" lastFinishedPulling="2026-03-14 10:04:01.828911834 +0000 UTC m=+3209.141522972" observedRunningTime="2026-03-14 10:04:02.286173042 +0000 UTC m=+3209.598784170" watchObservedRunningTime="2026-03-14 10:04:02.287771091 +0000 UTC m=+3209.600382209" Mar 14 10:04:03 crc kubenswrapper[4843]: I0314 10:04:03.278005 4843 generic.go:334] "Generic (PLEG): container finished" podID="52a5287e-5893-44d1-b5ab-ee535f4bec9a" containerID="89e033162395d140342b8dc679ed476e434560bbb88cefc21a58ee59156b4332" exitCode=0 Mar 14 10:04:03 crc kubenswrapper[4843]: I0314 10:04:03.278048 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558044-97x8m" event={"ID":"52a5287e-5893-44d1-b5ab-ee535f4bec9a","Type":"ContainerDied","Data":"89e033162395d140342b8dc679ed476e434560bbb88cefc21a58ee59156b4332"} Mar 14 10:04:04 crc kubenswrapper[4843]: I0314 10:04:04.572091 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:04 crc kubenswrapper[4843]: I0314 10:04:04.665114 4843 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc59p\" (UniqueName: \"kubernetes.io/projected/52a5287e-5893-44d1-b5ab-ee535f4bec9a-kube-api-access-sc59p\") pod \"52a5287e-5893-44d1-b5ab-ee535f4bec9a\" (UID: \"52a5287e-5893-44d1-b5ab-ee535f4bec9a\") " Mar 14 10:04:04 crc kubenswrapper[4843]: I0314 10:04:04.670852 4843 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a5287e-5893-44d1-b5ab-ee535f4bec9a-kube-api-access-sc59p" (OuterVolumeSpecName: "kube-api-access-sc59p") pod "52a5287e-5893-44d1-b5ab-ee535f4bec9a" (UID: "52a5287e-5893-44d1-b5ab-ee535f4bec9a"). InnerVolumeSpecName "kube-api-access-sc59p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 14 10:04:04 crc kubenswrapper[4843]: I0314 10:04:04.767341 4843 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc59p\" (UniqueName: \"kubernetes.io/projected/52a5287e-5893-44d1-b5ab-ee535f4bec9a-kube-api-access-sc59p\") on node \"crc\" DevicePath \"\"" Mar 14 10:04:05 crc kubenswrapper[4843]: I0314 10:04:05.295473 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558044-97x8m" event={"ID":"52a5287e-5893-44d1-b5ab-ee535f4bec9a","Type":"ContainerDied","Data":"31e6a27fd33c1170ff53bbf7cf4863af0588f9a2122fa79826f359cdbb86b720"} Mar 14 10:04:05 crc kubenswrapper[4843]: I0314 10:04:05.295525 4843 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558044-97x8m" Mar 14 10:04:05 crc kubenswrapper[4843]: I0314 10:04:05.295530 4843 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31e6a27fd33c1170ff53bbf7cf4863af0588f9a2122fa79826f359cdbb86b720" Mar 14 10:04:05 crc kubenswrapper[4843]: I0314 10:04:05.353173 4843 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558038-thp2w"] Mar 14 10:04:05 crc kubenswrapper[4843]: I0314 10:04:05.359726 4843 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558038-thp2w"] Mar 14 10:04:07 crc kubenswrapper[4843]: I0314 10:04:07.347889 4843 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6fea642-5dda-4baa-8b0e-9fe644a58a2f" path="/var/lib/kubelet/pods/c6fea642-5dda-4baa-8b0e-9fe644a58a2f/volumes" Mar 14 10:04:08 crc kubenswrapper[4843]: I0314 10:04:08.339811 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:04:08 crc kubenswrapper[4843]: E0314 10:04:08.340228 4843 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gwd22_openshift-machine-config-operator(8f55462c-aaf9-41fa-bae6-2f7fa118841b)\"" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" podUID="8f55462c-aaf9-41fa-bae6-2f7fa118841b" Mar 14 10:04:23 crc kubenswrapper[4843]: I0314 10:04:23.344896 4843 scope.go:117] "RemoveContainer" containerID="7d23d2e7a061750c735953bb62484a8e35f2be04005637e13e7060d7ef758ef6" Mar 14 10:04:24 crc kubenswrapper[4843]: I0314 10:04:24.463114 4843 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gwd22" event={"ID":"8f55462c-aaf9-41fa-bae6-2f7fa118841b","Type":"ContainerStarted","Data":"af075769944eb93ab5df2bcfc7f27b81e9393388b3dcc72b19cd2703ed68e52d"} Mar 14 10:04:59 crc kubenswrapper[4843]: I0314 10:04:59.564784 4843 scope.go:117] "RemoveContainer" containerID="4bb40ffe6dc4aa8b197798ffb07aef7c6e763f70ed866d6184b6c0e5ce130008" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515155231400024441 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015155231401017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015155222525016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015155222525015461 5ustar corecore